نوکیا نے خاموشی سے اپنا ایک اور اسمارٹ فون متعارف کرا دیا ہے۔
رواں سال یہ کمپنی نوکیا تھری، فائیو، سکس اور فلیگ شپ ڈیوائس 8 متعارف کراچکی تھی مگر اب اس نے نوکیا 7 بھی پیش کردیا ہے۔
دلچسپ بات یہ ہے کہ اس فون میں فلیگ شپ ڈیوائس نوکیا 8 کی طرح ڈوئل کیمرے اور بوتھی کیمرہ سافٹ ویئر دیا گیا ہے۔
مزید پڑھیں : نوکیا کا پہلا فلیگ شپ اینڈرائیڈ فون
اس فیچر کے تحت فرنٹ اور بیک کیمروں کو ایک ساتھ متحرک کرکے تصویر یا ویڈیو لی جاسکتی ہے، جبکہ ویڈیو کو فیس بک یا یوٹیوب پر لائیو نشر بھی کیا جاسکتا ہے۔
5.2 انچ کے نوکیا سیون میں کوالکوم اسنیپ ڈراگون 630 پراسیسر دیا گیا ہے جبکہ 4 جی بی ریم کے ساتھ 3000 ایم اے ایچ بیٹری ہے۔
اسی طرح 64 جی بی اسٹوریج ہے جس میں ایس ڈی کارڈ کے ذریعے مزید اضافہ کیا جاسکتا ہے جبکہ یہ فون سکس جی بی ریم اور 128 اسٹوریج کے آپشن کے ساتھ بھی دستیاب ہوگا۔
اس کے بیک پر ایک سولہ میگا پکسل جبکہ دوسرا پانچ میگا پکسل کیمرہ دیا گیا ہے۔
یہ بھی پڑھیں : نوکیا 3310 کی واپسی
اس کے فرنٹ پر پانچ میگا پکسل کیمرہ ہے اور نوکیا 8 کے اہم ترین فیچر کا حامل ہونے کے باوجود یہ قیمت میں اس سے 50 فیصد سستا ہے۔
اس فون کو چین میں متعارف کرایا گیا جہاں یہ چوبیس اکتوبر سے 2499 یوآن (لگ بھگ چالیس ہزار پاکستانی روپے) میں دستیاب ہوگا جبکہ نوکیا 8 کی قیمت 705 ڈالرز (75 ہزار پاکستانی روپے سے زائد) رکھی گئی تھی۔
دیگر ممالک میں یہ کب تک پیش کیا جائے گا، اس کے بارے میں کمپنی نے فی الحال کوئی اعلان نہیں کیا۔
CAPE CANAVERAL: Nasa astronauts took another spacewalk outside the International Space Station on Tuesday, this time to grease the robot arm’s new hand.
Commander Randy Bresnik ventured out for the second time in less than a week, along with Mark Vande Hei.
The pair replaced the latching mechanism on one end of the 58-foot robot arm on Thursday. The mechanism malfunctioned in August.
Tuesday’s work involved using a grease gun, which resembles a caulking gun, to keep the latching mechanism working smoothly. The two-part lube job is expected to spill into next week, in a third spacewalk.
These latches, or hands, are located on each end of the Canadian-built robot arm. They’re used to grab arriving US cargo ships and also allow the robot arm to move around the orbiting lab.
Launched in 2001 with the rest of the robot arm, the original latches were showing their age. Nasa plans to replace the latching mechanism on the opposite end of the arm early next year.
Published in Dawn, October 11th, 2017
Different methods of studying cosmic expansion yield slightly different results, including for the age of the universe. In a new study, astronomers from the Harvard-Smithsonian Center For Astrophysics have calculated that these discrepancies could be reconciled if the dark energy that drives cosmic acceleration were not constant in time.
The universe is not only expanding – it is accelerating outward, driven by what is commonly referred to as “dark energy.” The term is a poetic analogy to label for dark matter, the mysterious material that dominates the matter in the universe and that really is dark because it does not radiate light (it reveals itself via its gravitational influence on galaxies). Two explanations are commonly advanced to explain dark energy. The first, as Einstein once speculated, is that gravity itself causes objects to repel one another when they are far enough apart (he added this “cosmological constant” term to his equations). The second explanation hypothesizes (based on our current understanding of elementary particle physics) that the vacuum has properties that provide energy to the cosmos for expansion.
For several decades cosmologies have successfully used a relativistic equation with dark matter and dark energy to explain increasingly precise observations about the cosmic microwave background, the cosmological distribution of galaxies, and other large-scale cosmic features. But as the observations have improved, some apparent discrepancies have emerged. One of the most notable is the age of the universe: there is an almost 10% difference between measurements inferred from the Planck satellite data and those from so-called Baryon Acoustic Oscillation experiments. The former relies on far-infrared and submillimeter measurements of the cosmic microwave background and the latter on spatial distribution of visible galaxies.
CfA astronomer Daniel Eisenstein was a member of a large consortium of scientists who suggest that most of the difference between these two methods, which sample different components of the cosmic fabric, could be reconciled if the dark energy were not constant in time. The scientists apply sophisticated statistical techniques to the relevant cosmological datasets and conclude that if the dark energy term varied slightly as the universe expanded (though still subject to other constraints), it could explain the discrepancy. Direct evidence for such a variation would be a dramatic breakthrough, but so far has not been obtained. One of the team’s major new experiments, the Dark Energy Spectroscopic Instrument (DESI) Survey, could settle the matter. It will map over twenty-five million galaxies in the universe, reaching back to objects only a few billion years after the big bang, and should be completed sometime in the mid 2020’s.
Publication: Gong-Bo Zhao, et al., “Dynamical Dark Energy in Light of the Latest Observations,” Nature Astronomy 1, 627–632 (2017) doi:10.1038/s41550-017-0216-z
Source: Harvard-Smithsonian Center For Astrophysics
A team of engineers at MIT has developed a method for designing efficient computer chips may get miniature smart drones off the ground.
In recent years, engineers have worked to shrink drone technology, building flying prototypes that are the size of a bumblebee and loaded with even tinier sensors and cameras. Thus far, they have managed to miniaturize almost every part of a drone, except for the brains of the entire operation — the computer chip.
Standard computer chips for quadcoptors and other similarly sized drones process an enormous amount of streaming data from cameras and sensors, and interpret that data on the fly to autonomously direct a drone’s pitch, speed, and trajectory. To do so, these computers use between 10 and 30 watts of power, supplied by batteries that would weigh down a much smaller, bee-sized drone.
Now, engineers at MIT have taken a first step in designing a computer chip that uses a fraction of the power of larger drone computers and is tailored for a drone as small as a bottlecap. They will present a new methodology and design, which they call “Navion,” at the Robotics: Science and Systems conference, held this week at MIT.
The team, led by Sertac Karaman, the Class of 1948 Career Development Associate Professor of Aeronautics and Astronautics at MIT, and Vivienne Sze, an associate professor in MIT’s Department of Electrical Engineering and Computer Science, developed a low-power algorithm, in tandem with pared-down hardware, to create a specialized computer chip.
The key contribution of their work is a new approach for designing the chip hardware and the algorithms that run on the chip. “Traditionally, an algorithm is designed, and you throw it over to a hardware person to figure out how to map the algorithm to hardware,” Sze says. “But we found by designing the hardware and algorithms together, we can achieve more substantial power savings.”
“We are finding that this new approach to programming robots, which involves thinking about hardware and algorithms jointly, is key to scaling them down,” Karaman says.
The new chip processes streaming images at 20 frames per second and automatically carries out commands to adjust a drone’s orientation in space. The streamlined chip performs all these computations while using just below 2 watts of power — making it an order of magnitude more efficient than current drone-embedded chips.
Karaman, says the team’s design is the first step toward engineering “the smallest intelligent drone that can fly on its own.” He ultimately envisions disaster-response and search-and-rescue missions in which insect-sized drones flit in and out of tight spaces to examine a collapsed structure or look for trapped individuals. Karaman also foresees novel uses in consumer electronics.
“Imagine buying a bottlecap-sized drone that can integrate with your phone, and you can take it out and fit it in your palm,” he says. “If you lift your hand up a little, it would sense that, and start to fly around and film you. Then you open your hand again and it would land on your palm, and you could upload that video to your phone and share it with others.”
Karaman and Sze’s co-authors are graduate students Zhengdong Zhang and Amr Suleiman, and research scientist Luca Carlone.
From the ground up
Current minidrone prototypes are small enough to fit on a person’s fingertip and are extremely light, requiring only 1 watt of power to lift off from the ground. Their accompanying cameras and sensors use up an additional half a watt to operate.
“The missing piece is the computers — we can’t fit them in terms of size and power,” Karaman says. “We need to miniaturize the computers and make them low power.”
The group quickly realized that conventional chip design techniques would likely not produce a chip that was small enough and provided the required processing power to intelligently fly a small autonomous drone.
“As transistors have gotten smaller, there have been improvements in efficiency and speed, but that’s slowing down, and now we have to come up with specialized hardware to get improvements in efficiency,” Sze says.
The researchers decided to build a specialized chip from the ground up, developing algorithms to process data, and hardware to carry out that data-processing, in tandem.
Tweaking a formula
Specifically, the researchers made slight changes to an existing algorithm commonly used to determine a drone’s “ego-motion,” or awareness of its position in space. They then implemented various versions of the algorithm on a field-programmable gate array (FPGA), a very simple programmable chip. To formalize this process, they developed a method called iterative splitting co-design that could strike the right balance of achieving accuracy while reducing the power consumption and the number of gates.
A typical FPGA consists of hundreds of thousands of disconnected gates, which researchers can connect in desired patterns to create specialized computing elements. Reducing the number gates with co-design allowed the team to chose an FPGA chip with fewer gates, leading to substantial power savings.
“If we don’t need a certain logic or memory process, we don’t use them, and that saves a lot of power,” Karaman explains.
Each time the researchers tweaked the ego-motion algorithm, they mapped the version onto the FPGA’s gates and connected the chip to a circuit board. They then fed the chip data from a standard drone dataset — an accumulation of streaming images and accelerometer measurements from previous drone-flying experiments that had been carried out by others and made available to the robotics community.
“These experiments are also done in a motion-capture room, so you know exactly where the drone is, and we use all this information after the fact,” Karaman says.
For each version of the algorithm that was implemented on the FPGA chip, the researchers observed the amount of power that the chip consumed as it processed the incoming data and estimated its resulting position in space.
The team’s most efficient design processed images at 20 frames per second and accurately estimated the drone’s orientation in space, while consuming less than 2 watts of power.
The power savings came partly from modifications to the amount of memory stored in the chip. Sze and her colleagues found that they were able to shrink the amount of data that the algorithm needed to process, while still achieving the same outcome. As a result, the chip itself was able to store less data and consume less power.
“Memory is really expensive in terms of power,” Sze says. “Since we do on-the-fly computing, as soon as we receive any data on the chip, we try to do as much processing as possible so we can throw it out right away, which enables us to keep a very small amount of memory on the chip without accessing off-chip memory, which is much more expensive.”
In this way, the team was able to reduce the chip’s memory storage to 2 megabytes without using off-chip memory, compared to a typical embedded computer chip for drones, which uses off-chip memory on the order of a few gigabytes.
“Any which way you can reduce the power so you can reduce battery size or extend battery life, the better,” Sze says.
This summer, the team will mount the FPGA chip onto a drone to test its performance in flight. Ultimately, the team plans to implement the optimized algorithm on an application-specific integrated circuit, or ASIC, a more specialized hardware platform that allows engineers to design specific types of gates, directly onto the chip.
“We think we can get this down to just a few hundred milliwatts,” Karaman says. “With this platform, we can do all kinds of optimizations, which allows tremendous power savings.”
This research was supported, in part, by Air Force Office of Scientific Research and the National Science Foundation.
Source: Jennifer Chu, MIT News
A team of researchers has made the first demonstration of a solid state cooling device based on the electrocaloric effect. This thin flexible device that could keep smartphones and laptop computers cool and prevent overheating.
Engineers and scientists from the UCLA Henry Samueli School of Engineering and Applied Science and SRI International, a nonprofit research and development organization based in Menlo Park, California, have created a thin flexible device that could keep smartphones and laptop computers cool and prevent overheating.
The system’s flexibility also means it could eventually be used in wearable electronics, robotic systems and new types of personalized cooling systems. It is the first demonstration of a solid state cooling device based on the electrocaloric effect — a phenomenon in which a material’s temperature changes when an electric field is applied to it. The research was published September 15 in Science.
The method devised by UCLA and SRI researchers is very energy-efficient. It uses a thin polymer film that transfers heat from the heat source (a battery or processor, typically) to a “heat sink,” and alternates contact between the two by switching on and off the electric voltage. Because the polymer film is flexible, the system could be adapted for devices with complex curvature or with moving surfaces.
“We were motivated by the idea of devising a personalized cooling system,” said Qibing Pei, UCLA a professor of materials science and engineering and the study’s principal investigator. “For example, an active cooling pad could keep a person comfortable in a hot office and thus lower the electricity consumption for building air conditioning. Or it could be placed in a shoe insole or in a hat to keep a runner comfortable in the hot Southern California sun. It’s like a personal air conditioner.”
A major application could be in mobile and wearable electronics. As most smartphone and tablet users know, devices tend to heat up when they are used, particularly with power-intensive applications like video streaming. So although the devices are made with interior metal radiators designed to pull heat away from the battery and computer processors, they can still overheat, which can even cause them to shut down. And excessive heat can damage the devices’ components over time.
That tendency to overheat remains a major challenge for engineers, and with the anticipated introduction of more flexible electronic devices, it’s an issue that researchers and device manufacturers are working hard to address. The cooling systems in larger devices like air conditioners and refrigerators, which use a process called vapor compression, are simply too large for mobile electronics. (They’re also impractical for smartphones and wearable technology because they use a chemical coolant that is an environmental hazard.)
“The development of practical efficient cooling systems that do not use chemical coolants that are potent greenhouse gases is becoming even more important as developing nations increase their use of air conditioning,” said Roy Kornbluh, an SRI research engineer.
The UCLA–SRI system also has certain advantages over another advanced type of cooling system, called thermoelectric coolers, which require expensive ceramic materials and whose cooling capabilities don’t yet measure up to vapor compression systems.
Pei said the invention’s other potential applications could include being used in a flexible pad for treating injuries, or reducing thermal “noise” in thermographic cameras, which are used by scientists and firefighters, and in night-vision devices, among other uses.
The study’s lead authors are UCLA postdoctoral scholar Rujun Ma and doctoral student Ziyang Zhang, both members of Pei’s research group. Other authors are Kwing Tong, a UCLA graduate student; David Huber, a research engineer at SRI; and Yongho Sungtaek Ju, a UCLA professor of mechanical and aerospace engineering.
The research was supported by the Department of Energy’s Advanced Research Projects Agency–Energy and by the Air Force Office of Scientific Research. The researchers have submitted a U.S. patent application for the device.
Publication: Rujun Ma, et al., “Highly efficient electrocaloric cooling with electrostatic actuation,” Science 15 Sep 2017: Vol. 357, Issue 6356, pp. 1130-1134; DOI: 10.1126/science.aan5980
Source: Matthew Chin, UCLA News
By cutting off a process that cancerous cells rely, researchers from MIT have identified a possible new strategy for halting brain tumors.
MIT biologists have discovered a fundamental mechanism that helps brain tumors called glioblastomas grow aggressively. After blocking this mechanism in mice, the researchers were able to halt tumor growth.
The researchers also identified a genetic marker that could be used to predict which patients would most likely benefit from this type of treatment. Glioblastoma is usually treated with radiation and the chemotherapy drug temozolamide, which may extend patients’ lifespans but in most cases do not offer a cure.
“There are very few specific or targeted inhibitors that are used in the treatment of brain cancer. There’s really a dire need for new therapies and new ideas,” says Michael Hemann, an associate professor of biology at MIT, a member of MIT’s Koch Institute for Integrative Cancer Research, and a senior author of the study.
Drugs that block a key protein involved in the newly discovered process already exist, and at least one is in clinical trials to treat cancer. However, most of these inhibitors do not cross the blood-brain barrier, which separates the brain from circulating blood and prevents large molecules from entering the brain. The MIT team hopes to develop drugs that can cross this barrier, possibly by packaging them into nanoparticles.
The study, which appears in Cancer Cell on September 28, is a collaboration between the labs of Hemann; Jacqueline Lees, associate director of the Koch Institute and the Virginia and D.K. Ludwig Professor for Cancer Research; and Phillip Sharp, an MIT Institute Professor and member of the Koch Institute. The paper’s lead authors are former MIT postdoc Christian Braun, recent PhD recipient Monica Stanciu, and research scientist Paul Boutz.
Too much splicing
Several years ago, Stanciu and Braun came up with the idea to use a type of screen known as shRNA to seek genes involved in glioblastoma. This test involves using short strands of RNA to block the expression of specific genes. Using this approach, researchers can turn off thousands of different genes, one per tumor cell, and then measure the effects on cell survival.
One of the top hits from this screen was the gene for a protein called PRMT5. When this gene was turned off, tumor cells stopped growing. Previous studies had linked high levels of PRMT5 to cancer, but the protein is an enzyme that can act on hundreds of other proteins, so scientists weren’t sure exactly how it was stimulating cancer cell growth.
Further experiments in which the researchers analyzed other genes affected when PRMT5 was inhibited led them to hypothesize that PRMT5 was using a special type of gene splicing to stimulate tumor growth. Gene splicing is required to snip out portions of messenger RNA known as introns, that are not needed after the gene is copied into mRNA.
In 2015, Boutz and others in Sharp’s lab discovered that about 10 to 15 percent of human mRNA strands still have one to three “detained introns,” even though they are otherwise mature. Because of those introns, these mRNA molecules can’t leave the nucleus.
“What we think is that these strands are basically an mRNA reservoir. You have these unproductive isoforms sitting in the nucleus, and the only thing that keeps them from being translated is that one intron,” says Braun, who is now a physician-scientist at Ludwig Maximilian University of Munich.
In the new study, the researchers discovered that PRMT5 plays a key role in regulating this type of splicing. They speculate that neural stem cells utilize high levels of PRMT5 to guarantee efficient splicing and therefore expression of proliferation genes. “As the cells move toward their mature state, PRMT5 levels drop, detained intron levels rise, and those messenger RNAs associated with proliferation get stuck in the nucleus,” Lees says.
When brain cells become cancerous, PRMT5 levels are typically boosted and the splicing of proliferation-associated mRNA is improved, ultimately helping the cells to grow uncontrollably.
When the researchers blocked PRMT5 in tumor cells, they found that the cells stopped dividing and entered a dormant, nondividing state. PRMT5 inhibitors also halted growth of glioblastoma tumors implanted under the skin of mice, but they did not work as well in tumors located in the brain, because of the difficulties in crossing the blood-brain barrier.
Unlike many existing cancer treatments, the PRMT5 inhibitors did not appear to cause major side effects. The researchers believe this may be because mature cells are not as dependent as cancer cells on PRMT5 function.
The findings shed light on why researchers have previously found PRMT5 to be a promising potential target for cancer treatment, says Omar Abdel-Wahab, an assistant member in the Human Oncology and Pathogenesis Program at Memorial Sloan Kettering Cancer Center, who was not involved in the study.
“PRMT5 has a lot of roles, and until now, it has not been clear what is the pathway that is really important for its contributions to cancer,” says Abdel-Wahab. “What they have found is that one of the key contributions is in this RNA splicing mechanism, and furthermore, when RNA splicing is disrupted, that key pathway is disabled.”
The researchers also discovered a biomarker that could help identify patients who would be most likely to benefit from a PRMT5 inhibitor. This marker is a ratio of two proteins that act as co-factors for PRMT5’s splicing activity, and reveals whether PRMT5 in those tumor cells is involved in splicing or some other cell function.
“This becomes really important when you think about clinical trials, because if 50 percent or 25 percent of tumors are going to have some response and the others are not, you may not have a way to target it toward those patients that may have a particular benefit. The overall success of the trial may be damaged by lack of understanding of who’s going to respond,” Hemann says.
The MIT team is now looking into the potential role of PRMT5 in other types of cancer, including lung tumors. They also hope to identify other genes and proteins involved in the splicing process they discovered, which could also make good drug targets.
Spearheaded by students and postdocs from several different labs, this project offers a prime example of the spirit of collaboration and “scientific entrepreneurship” found at MIT and the Koch Institute, the researchers say.
“I think it really is a classic example of how MIT is a sort of bottom-up place,” Lees says. “Students and postdocs get excited about different ideas, and they sit in on each other’s seminars and hear interesting things and pull them together. It really is an amazing example of the creativity that young people at MIT have. They’re fearless.”
The research was funded by the Ludwig Center for Molecular Oncology at MIT, the Koch Institute Frontier Research Program through the Kathy and Curt Marble Cancer Research Fund, the National Institutes of Health, and the Koch Institute Support (core) Grant from the National Cancer Institute.
Publication: Christian J. Braun, et al., “Coordinated Splicing of Regulatory Detained Introns within Oncogenic Transcripts Creates an Exploitable Vulnerability in Malignant Glioma,” Cancer Cell, 2017; DOI:10.1016/j.ccell.2017.08.018
Source: Anne Trafton, MIT News