A serendipitous flythrough of the tail of a disintegrated comet has offered scientists a unique opportunity to study these remarkable structures.
Comet ATLAS fragmented just before its closest approach to the Sun last year, leaving its former tail trailing through space in the form of wispy clouds of dust and charged particles. The disintegration was observed by the Hubble Space Telescope in April 2020, but more recently the ESA spacecraft Solar Orbiter has flown close to the tail remnants in the course of its ongoing mission.
This lucky encounter has presented researchers with a unique opportunity to investigate the structure of an isolated cometary tail. Using combined measurements from all of Solar Orbiter’s in-situ instruments, the scientists have reconstructed the encounter with ATLAS’s tail. The resulting model indicates that the ambient interplanetary magnetic field carried by the solar wind ‘drapes’ around the comet, and surrounds a central tail region with a weaker magnetic field.
Source (Royal Astronomical Society, “Tail without a comet: the dusty remains of Comet ATLAS”, 19.07.2021)
An elegant new algorithm can significantly reduce the resource consumption of the world’s computer servers. Computer servers are as taxing on the climate as global air traffic combined, thereby making the green transition in IT an urgent matter. The researchers expect major IT companies to deploy the algorithm immediately.
Compared to existing load balancing methods, the new methodology can distribute efficiently clients among servers, ensuring that the load is balanced and the retrieval time is minimal. Mathematically, the improvement is equal to going from 100 steps to only 10. This method has already been applied by Vimeo, which stated that it has reduced the bandwidth usage by a factor of 8.
Source (University of Copenhagen – Faculty of Science. “Making computer servers worldwide more climate friendly.” ScienceDaily. ScienceDaily, 6 July 2021.)
Original paper: Aamand, A., Knudsen, J.B.T. and Thorup, M., 2021, June. Load balancing with dynamic set of balls and bins. In Proceedings of the 53rd Annual ACM SIGACT Symposium on Theory of Computing (pp. 1262-1275).
Researchers have developed a simple lab-based technique that allows them to look inside lithium-ion batteries and follow lithium ions moving in real time as the batteries charge and discharge, something which has not been possible until now.
The researchers from the University of Cambridge developed a low-cost technique based on optical microscopy called interferometric scattering microscopy to observe the insides of lithium-ion batteries. The method is able to observe how particles of lithium cobalt oxide (LCO) are charging or discharging by measuring the amount of scattered light. This will help to improve existing battery materials and accelerate their developments in the future.
“We found that there are different speed limits for lithium-ion batteries, depending on whether it’s charging or discharging,” said Dr Akshay Rao from the Cavendish Laboratory, who led the research. “When charging, the speed depends on how fast the lithium ions can pass through the particles of active material. When discharging, the speed depends on how fast the ions are inserted at the edges. If we can control these two mechanisms, it would enable lithium-ion batteries to charge much faster.”
Source (University of Cambridge. “Low-cost imaging technique shows how smartphone batteries could charge in minutes.” ScienceDaily. ScienceDaily, 23 June 2021.)
Merryweather, A.J., Schnedermann, C., Jacquet, Q., Grey, C.P. and Rao, A., 2021. Operando optical tracking of single-particle ion dynamics in batteries. Nature, 594(7864), pp.522-528.
Nano-sized particles have been engineered in a new way to improve detection of tumors within the body and in biopsy tissue, a research team reports. The advance could enable identifying early stage tumors with lower doses of radiation.
Researchers from the KTH Royal Institute of Technology have developed “core-shell nanoparticles” which may be used in the future for targeted diagnostics, instead of current methods that use optical or X-ray fluorescence contrast agents. The tests performed in the laboratory on mice have shown that the new particles are able to detect early-stage tumours of only a few millimetres in size. “Nanoparticles of different size, originating from the same material, don’t appear to be distributed in the blood in the same concentrations,” Muhammet Toprak, Professor of Materials Chemistry at KTH, says. “That’s because when they come into contact with your body, they’re quickly wrapped in various biological molecules — which gives them a new identity.”
Source (KTH, Royal Institute of Technology. “New nanoparticle design paves way for improved detection of tumors.” ScienceDaily. ScienceDaily, 2 June 2021.)
Original paper: Saladino, G.M., Vogt, C., Li, Y., Shaker, K., Brodin, B., Svenda, M., Hertz, H.M. and Toprak, M.S., 2021. Optical and X-ray Fluorescent Nanoparticles for Dual Mode Bioimaging. ACS nano, 15(3), pp.5077-5085.
Researchers have developed a method for producing looping videos from one image. The technique is specialized in fluid motion seen in water, smoke or clouds. After training a deep neural network on thousands of images, the framework is capable of estimating motion. The solution posed several challenges, from which the most difficult to overcome was the employment of the splatting technique. Through it, holes in the top part of the images appeared. Using the previously obtained motion, the authors created a symmetric splatting methodology that merges the flow bidirectionally. A presentation of the published work can be seen in the following video.
For an in-depth understanding of the paper, please see the following resources:
Source (University of Washington. “Researchers can turn a single photo into a video.” ScienceDaily. ScienceDaily, 15 June 2021.)
Original paper: Holynski, A., Curless, B., Seitz, S.M. and Szeliski, R., 2020. Animating Pictures with Eulerian Motion Fields. arXiv preprint arXiv:2011.15128.
Dark matter is an elusive substance that makes up 80% of the universe. It also provides the skeleton for what cosmologists call the cosmic web, the large-scale structure of the universe that, due to its gravitational influence, dictates the motion of galaxies and other cosmic material. However, the distribution of local dark matter is currently unknown because it cannot be measured directly. Researchers must instead infer its distribution based on its gravitational influence on other objects in the universe, like galaxies.
Previous attempts to map the cosmic web started with a model of the early universe and then simulated the evolution of the model over billions of years. However, this method is computationally intensive and so far has not been able to produce results detailed enough to see the local universe. In the new study, the researchers took a completely different approach, using machine learning to build a model that uses information about the distribution and motion of galaxies to predict the distribution of dark matter.
The researchers built and trained their model using a large set of galaxy simulations, called Illustris-TNG, which includes galaxies, gasses, other visible matter, as well as dark matter. The team specifically selected simulated galaxies comparable to those in the Milky Way and ultimately identified which properties of galaxies are needed to predict the dark matter distribution. The research team then applied their model to real data from the local universe from the Cosmicflow-3 galaxy catalog. The catalog contains comprehensive data about the distribution and movement of more than 17 thousand galaxies in the vicinity of the Milky Way — within 200 megaparsecs.
The map successively reproduced known prominent structures in the local universe, including the “local sheet” — a region of space containing the Milky Way, nearby galaxies in the “local group,” and galaxies in the Virgo cluster — and the “local void” — a relatively empty region of space next to the local group. Additionally, it identified several new structures that require further investigation, including smaller filamentary structures that connect galaxies. For example, it has been suggested that the Milky Way and Andromeda galaxies may be slowly moving toward each other, but whether they may collide in many billions of years remains unclear. Studying the dark matter filaments connecting the two galaxies could provide important insights into their future.
Adapted and abridged from Source (Penn State. “Dark matter map reveals hidden bridges between galaxies.” ScienceDaily. ScienceDaily, 25 May 2021.)
Original paper: Sungwook E. Hong, Donghui Jeong, Ho Seong Hwang, Juhan Kim. Revealing the Local Cosmic Web from Galaxies by Deep Learning. The Astrophysical Journal, 2021; 913 (1): 76 DOI: 10.3847/1538-4357/abf040
The researchers at MIT (Massachusetts Institute of Technology) have developed a robot that uses radio waves, which can pass through walls, to sense occluded objects. The robot, called RF-Grasp, combines this powerful sensing with more traditional computer vision to locate and grasp items that might otherwise be blocked from view. The advance could one day streamline e-commerce fulfillment in warehouses or help a machine pluck a screwdriver from a jumbled toolkit.
Using optical vision alone, robots can’t perceive the presence of an item packed away in a box or hidden behind another object on the shelf — visible light waves, of course, don’t pass through walls. But radio waves can. For decades, radio frequency (RF) identification has been used to track everything from library books to pets. RF identification systems have two main components: a reader and a tag. The tag is a tiny computer chip that gets attached to — or, in the case of pets, implanted in — the item to be tracked. The reader then emits an RF signal, which gets modulated by the tag and reflected back to the reader.
The reflected signal provides information about the location and identity of the tagged item. The technology has gained popularity in retail supply chains — Japan aims to use RF tracking for nearly all retail purchases in a matter of years. The researchers realized this profusion of RF could be a boon for robots, giving them another mode of perception.
RF Grasp uses both a camera and an RF reader to find and grab tagged objects, even when they’re fully blocked from the camera’s view. It consists of a robotic arm attached to a grasping hand. The camera sits on the robot’s wrist. The RF reader stands independent of the robot and relays tracking information to the robot’s control algorithm. So, the robot is constantly collecting both RF tracking data and a visual picture of its surroundings. Integrating these two data streams into the robot’s decision making was one of the biggest challenges the researchers faced.
Adapted and abridged from Source (Massachusetts Institute of Technology. “A robot that senses hidden objects: System uses penetrative radio frequency to pinpoint items, even when they’re hidden from view.” ScienceDaily. ScienceDaily, 1 April 2021.)
Original paper: Boroushaki, T., Leng, J., Clester, I., Rodriguez, A. and Adib, F., 2020. Robotic Grasping of Fully-Occluded Objects using RF Perception. arXiv preprint arXiv:2012.15436.
Three out of four bites of food we eat depend on pollination. Bumblebees, solitary bees, hoverflies, butterflies, moths, wasps, beetles, and flies are all essential to keeping nature healthy. However, the pollinators are in serious decline. Around four in five crop and wild flowering plant species in the EU depend on animal pollination.
The European Commission has teamed up with renowned architect Vincent Callebaut to create the futuristic Pollinator Park. In collaboration with world renowned ‘archiobiotect’ Vincent Callebaut, you are invited into a 30-minute interactive and emotionally engaging virtual reality experience that immerses you in a futuristic world where man and nature co-exist in harmony, hoping to change your perspective and help turn the tide.
The interactive experience is set in 2050 where a cascade of ecological crises has impoverished the world and pollinating insects have all but disappeared. Visitors can walk through different steps in a futuristic farm, which provides a safe haven for pollinators and is an eye-opener for visitors.
Adapted and abridged from Source (This VR experience imagines a future without vital pollinators, Euronews, 30.04.2021)
Researchers at the University of California San Diego have recently developed a theoretical model that describes how the ocean, the wind and the birds in flight interact. UC San Diego mechanical engineering Ph.D. student Ian Stokes and adviser Professor Drew Lucas, of UC San Diego’s Department of Mechanical and Aerospace Engineering and Scripps Institution of Oceanography, found that pelicans can completely offset the energy they expend in flight by exploiting wind updrafts generated by waves through what is known as wave-slope soaring. In short, by practicing this behavior, sea-birds take advantage of winds generated by breaking waves to stay aloft.
The model could be used to develop better algorithms to control drones that need to fly over water for long periods of time, the researchers said. Potential uses do not stop there. The model can also serve as a basic prediction for the winds generated by passing swell, which is important to physicists that study how the ocean and atmosphere interact in order to improve weather forecasting.
Adapted and abridged from Source (University of California – San Diego. “The wave beneath their wings: Researchers work out intricate dance between waves, wind, and gliding pelicans.” ScienceDaily. ScienceDaily, 21 April 2021.)
Original paper: Stokes, I.A. and Lucas, A.J., 2021. Wave-slope soaring of the brown pelican. Movement Ecology, 9(1), pp.1-13.