RF & Microwave Lab.     (國立中山大學 射頻與微波實驗室)
We Connect People Together!
/  討論區主頁
   /  射頻與微波發布區
      /  雷達科技發展與應用
無發表權
到 ( 1 | 2 下頁 )
發布者內容列
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 雷達科技發展與應用

Seeing Through Walls

Youtube
http://www.youtube.com/watch?v=H5xmo7iJ7KA
(USA, FMCW Technology)

Seeing Through Walls
October 19, 2011
Researchers at MIT's Lincoln Lab have developed new radar technology that provides real-time video of what's going on behind solid walls.
The ability to see through walls is no longer the stuff of science fiction, thanks to new radar technology developed at MIT's Lincoln Laboratory.
Much as humans and other animals see via waves of visible light that bounce off objects and then strike our eyes' retinas, radar "sees" by sending out radio waves that bounce off targets and return to the radar's receivers. But just as light can't pass through solid objects in quantities large enough for the eye to detect, it's hard to build radar that can penetrate walls well enough to show what's happening behind. Now, Lincoln Lab researchers have built a system that can see through walls from some distance away, giving an instantaneous picture of the activity on the other side.
The researchers' device is an unassuming array of antenna arranged into two rows — eight receiving elements on top, 13 transmitting ones below — and some computing equipment, all mounted onto a movable cart. But it has powerful implications for military operations, especially "urban combat situations," says Gregory Charvat, technical staff at Lincoln Lab and the leader of the project.
Waves through walls
Walls, by definition, are solid, and that's certainly true of the four- and eight-inch-thick concrete walls on which the researchers tested their system.
At first, their radar functions as any other: Transmitters emit waves of a certain frequency in the direction of the target. But in this case, each time the waves hit the wall, the concrete blocks more than 99 percent of them from passing through. And that's only half the battle: Once the waves bounce off any targets, they must pass back through the wall to reach the radar's receivers — and again, 99 percent don't make it. By the time it hits the receivers, the signal is reduced to about 0.0025 percent of its original strength.
But according to Charvat, signal loss from the wall is not even the main challenge. "[Signal] amplifiers are cheap," he says. What has been difficult for through-wall radar systems is achieving the speed, resolution and range necessary to be useful in real time. "If you're in a high-risk combat situation, you don't want one image every 20 minutes, and you don't want to have to stand right next to a potentially dangerous building," Charvat says.
The Lincoln Lab team's system may be used at a range of up to 60 feet away from the wall. (Demos were done at 20 feet, which Charvat says is realistic for an urban combat situation.) And, it gives a real-time picture of movement behind the wall in the form of a video at the rate of 10.8 frames per second.
Filtering for frequencies
One consideration for through-wall radar, Charvat says, is what radio wavelength to use. Longer wavelengths are better able to pass through the wall and back, which makes for a stronger signal; however, they also require a correspondingly larger radar apparatus to resolve individual human targets. The researchers settled on S-band waves, which have about the same wavelength as wireless Internet — that is, fairly short. That means more signal loss — hence the need for amplifiers — but the actual radar device can be kept to about eight and a half feet long. "This, we believe, was a sweet spot because we think it would be mounted on a vehicle of some kind," Charvat says.
Even when the signal-strength problem is addressed with amplifiers, the wall — whether it's concrete, adobe or any other solid substance — will always show up as the brightest spot by far. To get around this problem, the researchers use an analog crystal filter, which exploits frequency differences between the modulated waves bouncing off the wall and those coming from the target. "So if the wall is 20 feet away, let's say, it shows up as a 20-kilohertz sine wave. If you, behind the wall, are 30 feet away, maybe you'll show up as a 30-kilohertz sine wave," Charvat says. The filter can be set to allow only waves in the range of 30 kilohertz to pass through to the receivers, effectively deleting the wall from the image so that it doesn't overpower the receiver.
"It's a very capable system mainly because of its real-time imaging capability," says Robert Burkholder, a research professor in Ohio State University's Department of Electrical and Computer Engineering who was not involved with this work. "It also gives very good resolution, due to digital processing and advanced algorithms for image processing. It's a little bit large and bulky for someone to take out in the field," he says, but agrees that mounting it on a truck would be appropriate and useful.
Monitoring movement
In a recent demonstration, Charvat and his colleagues, Lincoln Lab assistant staff John Peabody and former Lincoln Lab technical staff Tyler Ralston, showed how the radar was able to image two humans moving behind solid concrete and cinder-block walls, as well as a human swinging a metal pole in free space. The project won best paper at a recent conference, the 2010 Tri-Services Radar Symposium.
Because the processor uses a subtraction method — comparing each new picture to the last, and seeing what's changed — the radar can only detect moving targets, not inanimate objects such as furniture. Still, even a human trying to stand still moves slightly, and the system can detect these small movements to display that human's location.
The system digitizes the signals it receives into video. Currently, humans show up as "blobs" that move about the screen in a bird's-eye-view perspective, as if the viewer were standing on the wall and looking down at the scene behind. The researchers are currently working on algorithms that will automatically convert a blob into a clean symbol to make the system more end-user friendly. "To understand the blobs requires a lot of extra training," Charvat says.
With further refinement, the radar could be used domestically by emergency-response teams and others, but the researchers say they developed the technology primarily with military applications in mind. Charvat says, "This is meant for the urban war fighter … those situations where it's very stressful and it'd be great to know what's behind that wall."
SOURCE: Massachusetts Institute of Technology


_________________

 2011-10-25 21:04個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 Re: Seeing Through Walls

More See-Through-Wall Videos


UK: Prism 200 system, UWB technology
http://www.youtube.com/watch?v=t2ehrcyAnXc


Japan: Through wall scope, UWB technology
http://www.youtube.com/watch?NR=1&v=i2F-xcoY2Rg

Israel: XAVER 400/800, UWB technology
http://www.youtube.com/watch?v=zh5da2y4JME
http://www.camero-tech.com/product.php?ID=39
http://www.camero-tech.com/product.php?ID=40


More information:
http://www.dailymail.co.uk/sciencetech/article-1382944/Back-pack-future-Prism-200c-device-lets-walls.html


_________________

 2013-01-09 23:11個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 International Success For Novel Norwegian Radar Technology

RF Globalnet, November 11, 2013

Mine clearing, monitoring of vital signs and robot vision are among the many potential applications for Novelda’s innovative radar technology. The company’s tiny, unique processor chips are at the core of it all.

Novelda’s hard-earned success began with a project on fundamental radar technology using principles developed at the University of Oslo. With the knowledge acquired, the Norwegian company has developed chips that are extremely fast, highly precise, run on very low power, and can be placed directly onto objects rather than metres away – qualities that are making Novelda stand out on the global market.

The company is now launching its new generation of impulse-based radar chips with ultra-high resolution.

Long-term investment leads to commercial products
Novelda’s radar technology can be applied to product development within a wide variety of sensor application areas. It has taken a lot of effort to transform the start-up from 2004 into the cutting-edge technology leader it is today.

“Much about our technology is completely new, and it takes lots of long-term planning and methodical work to develop attractive commercial products based on this technology,” explains Alf-Egil Bogen, CEO of Novelda AS. “The funding we have received all along the way under the Research Council of Norway’s programmes has in turn triggered private investment. This combined funding has been critical, providing us with opportunities we otherwise never would have had for carrying out the necessary R&D.”

The Research Council’s programme for User-driven Research-based Innovation (BIA) has provided funding support for three Novelda projects.

“The first of our projects with funding under the BIA programme ran from 2006 to 2008 and focused on the fundamental radar technology,” recounts Dag T. Wisland, president and co-founder of the company. “Only now have the results truly come together, and we are launching our first proprietary commercial sensor. This illustrates how crucial it is to take a long-term approach when bringing such a new technology to the commercial market.”

In addition to the BIA programme, Novelda has also received support from the Research Council’s large-scale programme on Core Competence and Value Creation in ICT (VERDIKT) and the EU Commission under EUREKA’s R&D programme, Eurostars.

Practical uses for Novelda technology
Customers all around the world are using Novelda’s technology to develop applications such as measuring asphalt thickness and snow depth. Another example is the collaboration between Novelda and Samsung Thales launched in April to develop a landmine detector mounted on vehicles. Tests so far are very promising.

“It’s very gratifying for us at Novelda to supply technology that can help to save lives by finding landmines,” says Mr Bogen.

“Through projects like these,” adds company CMO Aage Kalsæg, “we learn about new kinds of application areas, and at the same time our partners learn about our technology and how it can be used to develop other applications. That combination opens a lot of exciting doors.”

Potential medical applications
There is also widespread use for Novelda’s radar transceivers in the health care sector, where they can be used in applications such as monitoring heart rate, taking wireless ECG readings, and measuring fluid in the lungs. These were targeted in the company’s second BIA-funded project, which ran from 2009 to 2011 and included cooperation with the University of Oslo.

“We get lots of attention from cardiologists who are amazed by what we can do with our radar technology,” says CEO Bogen. “Health care is one of the areas with the greatest potential for us, but this is where it takes the most time and effort to come out with a finished, profitable product. We are in the process of working out a business plan and looking at various medical projects to go forward with.”

Working their way up the value chain
The third Novelda project with funding under the BIA programme started up this summer. It is a collaboration with the Norwegian University of Science and Technology (NTNU) in Trondheim.

“In this project we’re looking at how we can keep developing not only the actual radar but also the sensor that processes the radar image,” explains Mr Wisland. “This will enable us to move up the value chain, fill in the gap between the technology and the customer, and bring us closer to the market.”

“The solution we are working towards, a mini-radar, will make the technology available even to markets with no prior involvement with radar expertise,” adds CMO Kalsæg, “so customers will be able to further develop their own applications and products.”

Radar with Superman vision
Novelda’s mini-radar will open doors to more and more new application areas, such as building automation, security surveillance, technology for “intelligent homes” and more.
“Our radar can see right through things, so it could be placed out of sight, such as inside a wall,” explains Mr Bogen. “It is also very accurate and easy to program to define its range of detection, which makes it ideal for surveillance. It could also control an air-conditioning system, for instance, or count the people in a room, or be used as a parking sensor.”

“Helicopter technology is another interesting area,” says Mr Kalsæg. “Many unmanned helicopter drones have trouble landing gently in rugged terrain. Conventional sensors see the top of high grass and calibrate for landing on it, and then hit the ground harder. Our radar sees through the grass to the actual ground, facilitating a soft landing.”

Robot vision and consumer technology
Novelda is also working to develop robot vision as an application area. A robot using the mini-radar could avoid physical obstacles and navigate optimally by sensing its surroundings, regardless of light conditions. It would also be able to see through curtains and other such light objects. Ultimately, Novelda’s developers hope to make the radar capable of mapping a room and laying out its own route to follow.

“The next breakthrough,” Mr Bogen predicts, “will be in consumer-oriented radars in televisions, smoke detectors, thermostats, telephones and portable computer devices.”

The current project on the mini-radar focuses in large part on determining the proper future course for the company’s products. “The NTNU people are helping us to assess user experience and time to market,” says Mr Wisland. “They are also involved in the innovation processes and are working on product design as well.”


_________________

 2015-08-24 16:51個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 Robots See Through Walls Using Wi-Fi

RF Globalnet. August 5, 2014



Now You Can See The Invisible

Researchers at UCSB enable robots to see through solid walls with Wi-Fi

Wi-Fi makes all kinds of things possible. We can send and receive messages, make phone calls, browse the Internet, even play games with people who are miles away, all without the cords and wires to tie us down. At UC Santa Barbara, researchers are now using this versatile, everyday signal to do something different and powerful: looking through solid walls and seeing every square inch of what’s on the other side. Built into robots, the technology has far-reaching possibilities.

“This is an exciting time to be doing this kind of research,” said Yasamin Mostofi, professor of electrical and computer engineering at UCSB. For the past few years, she and her team have been busy realizing this X-ray vision, enabling robots to see objects and humans behind thick walls through the use of radio frequency signals. The patented technology allows users to see the space on the other side and identify not only the presence of occluded objects, but also their position and geometry, without any prior knowledge of the area. Additionally, it has the potential to classify the material type of each occluded object such as human, metal or wood.

The combination of imaging technology and automated mobility can make these robots useful in situations where human access is difficult or risky, and the ability to determine what is in a given occluded area is important, such as search and rescue operations for natural or man-made disasters.

But the technology is not limited to robots; it can be implemented on a Wi-Fi-enabled gadget or a Wi-Fi network. Built into an existing network the technology can be used to monitor the presence and location of objects and people throughout a built space, which opens possibilities for catching intruders, or watching over the elderly. It can also provide information for smart building applications to optimize services that depend on the level of occupancy of a building, such as heating and cooling. Further developed, the technology may even be useful in preliminary body scan and health monitoring via a WiFi-enabled handheld device — a real-life Star Trek tricorder.

For more information, please visit the project's website:
http://www.ece.ucsb.edu/~ymostofi/SeeThroughImaging.html

SOURCE: The Regents of the University of California


_________________

 2015-08-24 17:31個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 地表下3D成像技術成功探測地雷

電子工程專輯 2015年07月21日

英國巴斯大學(University of Bath)的研究人員利用基於電容響應和平面磁感應的雙重斷層攝影技術,可望開發出能夠探測地雷的全新3D攝影系統。由該校工程斷層攝影實驗室(ETL)教授Manuchehr Soleimani所主導的GBP 100,000三年期研究計劃,目的在於提供一種能在深達10公分地底的多變地形中,為單個爆炸裝置區別塑料影像與金屬影像的技術。

Soleimani的實驗室參與由英國工程暨物理科學研究委員會(EPSRC)舉辦的一項競賽,贏得了Bobby Charlton爵士慈善團體Find A Better Way (FABW)的資金贊助。

儘管像雷達或金屬探測器等「透地雷達」探測技術可能更成熟,但為了增加經濟效益以及易於隱藏,製造商在製造具人員殺傷力的地雷時已經改用塑料以取代金屬,使得地雷探測過程越來越困難。

因此,目前需要能探測金屬以外的材料。這一直是Dr. Soleimani最近發表的「電力與電磁斷層影像與立體影像建構」關注重點。在「平面陣列3D電容斷層攝影」與「3D近表面下成像的平面磁感應斷層攝影」這兩份研究報告中,Soleimani展示可觀的成果,他表示可望用於設計先進多模態電容/電感攝影新技術。

針對電容斷層攝影(ECT)部份,Soleimani的研究團隊利用一組佈置在4x3矩陣陣列(總面積約250x250mm,厚度為4mm)上的12個銅電極,以及一個12通道的電容測量儀器,拍攝放置在感測器陣列前方物件的介電常數特性。研究人員能夠以僅超過感測器陣列長度一半的深度拍攝電介質固體影像。

而在磁感應斷層攝影(MIT)部份,感應線圈以及渦流則被用於3D映射探測物件的被動電磁特性。在此的感測器以與軸心垂直的圓形方式放置。根據序列線圈激勵和響應分析,重新建構3D影像。

「我們正在尋找整合這兩種感測器類型的方法,利用正申請中的專利技術,讓我們能從一種感測器類型中取得3D影??像訊息,並從另一感測器取得提示資訊,從而完成畫面,」Soleimani表示。

「我們的想法是在整個頻率範圍中同時使用成像技術,並取得被成像物質的光譜分析戳記,」他說。

「實際上,我們可以打造一系列的材料戳記資料庫,讓電介質材料埋層影像能夠產生精確的組合。根據其形狀與組合物,這可用於關聯已知地雷的搭配分佈表,從而加速查找確認過程,」他解釋。

不過,研究人員還需要進一步的研究,才能分析周遭可能埋設地雷的地面特性,讓感測器能夠更有效地區別爆炸物。Soleimani對此相當樂觀,他表示,研究人員們已經利用粗略模型取得了很人振奮的結果。

此外,研究人員還必須在感測器陣列大小、複雜度與成像速度之間,取得權衡折衷。也許有人想要在一次掃描中涵蓋相當大的區域,那可能得採用安裝在火星機器手臂上的感測器陣列吧!


_________________

 2015-08-27 15:35個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 MIT最新看穿牆技術-RF Capture

美國研究員成功以Wi-Fi偵測牆後人身形及動作

Unwire HK, 2015年10月30日

隔牆偵測牆後的人,在不少間諜電影中都經常出現,而且這種技術其實早就存在。隨著不斷改良之後,最近美國麻省理工學院(MIT)的研究人員更成功製作出一款名為「RF-Capture」的裝置,可以透過將 Wi-Fi 訊號射穿牆,然後再偵測到牆後人物的身形及動作,據知準確度更高達 90%。

其實早在 2013 年,MIT 電腦科學及人工智慧實驗室的研究小組就曾經發展過類似的技術,透過無線電頻率技術去偵測牆或障礙物另一邊的舉動。不過這次 RF-Capture 的技術就更加純粹及複雜,能夠仔細辨識到目標身形的細微差異,據了解在一項測試中成功分辨到 15 個不同的人,準確率高達 90%,而且甚至更可判斷一個人的呼吸模式及心跳率,比起單純偵測到動作更加厲害。

至於 RF-Capture 的運動原理其他不算太過複雜,首先裝置會發射出 Wi-Fi 訊號,當訊號穿過牆並接觸到目標人物,接著便會反射回來。之後 RF-Capture 便會接收這些反彈回來的訊號,而由於不同身體部位所反彈的訊號強弱會有所不同,所以當將所有訊號集合起來並進行運算,最後就可以組合出目標人物的輪廓(就好像看剪影一樣)。據研究人員顯示,有關技術的應用將有很多可能性,比如可以追蹤獨居長者的活動,確定他們是否發生意外;又或者能應用在智慧家居之上,讓使用者即使隔住牆壁,都可以用特定的手勢去開關各種家電。


無線X光透視技術能隔牆認人

電子工程專輯: 2015年11月04日

美國麻省理工學院(MIT)電腦暨人工智慧實驗室(CSAIL)的研究人員最近開發出一種可利用無線電訊號隔牆追蹤人體身形與動作的RF-Capture系統。

這套RF-Capture系統能讓相同頻率範圍的訊號(即Wi-Fi)隔著一道厚牆定位出室內有多少人、根據人的身形輪廓來區別與辨識不同的人,以及追蹤其身體動作與四肢的位置。

目前雖然也有其他追蹤人體動作的方式,但可能不適用於一些特定情況。例如,微軟(Microsoft)的Kinect系統靠外線來追蹤遊戲的動作。它要靠直視的視線作業,因此會被其間的障礙物所阻礙。同樣地,傳統的電影動態擷取技術有賴於固定在主體上的感測器,因而也不實用。

再來就是以雷達為基礎的系統,例如使用毫米波(mmWave)或次毫米波的機場安檢掃描儀。其中有些儀器甚至還能精準到成像人體骨骼,但所費不貲且體積相當龐大,只能用於短距離內,也無法因應障礙物的情況。

還有像Camero公司採用厘米波(3-30GHz)的Xaver系列產品也是以雷達系統為基礎,但這些系統是專為軍方與執法行動提供封閉空間的情勢感知而設計的。雖然也可能適用於辨識人在室內的位置,但研究人員表示,這種「穿牆成像系統」無法準確地追蹤人的肢體動作或建構出身體輪廓。

根據研究人員——MIT教授Dina Katabi、博士研究生Fredo Durand與Chen-Yu Hsu,以及實習生Hongzi Mao表示,他們開發出的系統能以98%的高度準確性區別兩個不一樣的人。而當室內有10個人的情況下,這一準確度則降至92%,15個人時的準確度則為88%。

此外,研究人員指出,RF-Capture系統還能以92%的準確度追蹤肢體動作,也能像Kinect一樣追蹤體感動作。

該技術已經準備上市了,它將會是一款名為Emerald的監測器,利用無線電波追蹤屋內的幾個房間,偵測年長者或生病的人是否發生跌倒或面臨需要協助的情況。

此外,紐約市警察局長久以來使用配備X光線掃描儀的警車,進行隨機炸彈搜尋的任務。如今,在RF-Capture系統進一步強化後,可望用於追捕這些炸彈攻擊或相關執法行動,當然各種商業與企業的偵測應用就更不用說了。



_________________

 2015-11-04 12:04個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 Wi-Fi-Based Sensing: The Myths, The Facts, The Opportunities!

RF Globalnet, September 1, 2016

Last year, Laser Interferometer Gravitational-Wave Observatory (LIGO) researchers confirmed Albert Einstein’s theory about the presence of gravitational waves. Using the most stable optical system ever built, they measured a gravitational wave's low-frequency effect through distance changes of about 10-18 m, about 1/1,000 the diameter of a proton. Now, researchers are applying that same principle — using ultrasensitive equipment to capture minute phenomena — to Wi-Fi.

We all know that a wireless router emits Wi-Fi signals and receives information back about those signals in real-time, making adjustments on the fly to better communicate with the devices accessing the signal. But when something moves within that wireless signal’s coverage, there is a slight change in the signal’s frequency, known as the Doppler frequency shift. While these shifts are very small — often just a few hertz within Wi-Fi signals operating at a 20 MHz bandwidth — researchers and entrepreneurs have developed both the technology to detect the shifts and the algorithms to interpret their meaning.

University of Washington computer scientists leveraged this knowledge in 2013 to create WiSee, gesture-recognition technology they envisioned being used to control entire smart homes through simple motions (e.g., waves of the hand). No additional sensors, remotes, RFID implants, or wearables were necessary, and the user did not have to be in the same room as the receiver.

That same year, Massachusetts Institute of Technology (MIT) researchers discovered that they could track people’s movements through walls by analyzing the reflected Wi-Fi signals.

Recently, there has been a reemergence of interest — and skepticism and conspiracy theories —surrounding this Wi-Fi-based sensing. Last month, computer science researchers at Northwestern Polytechnical University in China published a paper detailing a system that, reading changes in Wi-Fi signals, can identify humans as they walk through a door with about 90 percent accuracy. However, this system must first learn the subjects’ body shapes to properly identify them later.

More recent MIT efforts have resulted in Wi-Fi technologies that can track breathing and heartbeat, as well as a system being pitched as a home monitor — designed to keep tabs on the well-being of elderly or developmentally disabled family members. Similarly, an assistant professor of electrical and computer engineering at Michigan State University’s College of Engineering has received a two-year, $171,600 NSF grant to continue developing his adaptation of Wi-Fi as a sensing technology. Like the MIT researchers, he is attempting to create a system that can alert family or caregivers of abnormal events in a loved one’s home, minus the intrusiveness of wearable devices or cameras.

In 2014, technology was developed by a Berkeley Ph.D. student-led team that can “hear” what someone is saying by analyzing the effect of their mouth’s movements on the Wi-Fi signal. In the same vein, a Michigan State University Ph.D. student led the research behind a system (introduced last year) that can track keystrokes by – you guessed it – analyzing Doppler frequency shift. Both systems have prompted a number of reactionary articles fearful of Wi-Fi routers spying on us.

However, most of those articles fail to account for the fact that these experiments were conducted under controlled conditions, by willing participants. Add more voices or unnecessary movements to the equation, and the accuracy of the results suffers considerably. Kamran Ali, who led the MSU research, told The Atlantic that these limitations mean such technology is “not a big privacy concern for now.”

Still, developers would be remiss if they did not pursue such applications. It’s foolhardy to believe that others aren’t striving to perfect what could be groundbreaking surveillance technology. Routers themselves could be embedded with movement-tracking receivers, eliminating the need for third-party devices. More advanced receivers will be developed, capable of detecting even subtler shifts in frequency and more effectively ignoring interference (RF engineers are rather adept at filtering out noise), and researchers will develop algorithms that more accurately interpret those signal disturbances.

While acknowledging that military and defense applications will be vetted for just about any new technology, I’d prefer to muse over the consumer-driven applications of Wi-Fi sensing, such as the aforementioned researchers’ efforts to create a home monitoring system. I’m also still waiting for a Wi-Fi-based gesture-recognition system that can control the lights in my house, or adjust the volume on my TV.

Consider, also, quality-of-life solutions. A Wi-Fi sensing receiver could be trained to accurately interpret sign language, allowing for new possibilities in communication —picture a police station or emergency room where no sign language interpreter is present and fast, effective communication is vital. Continuing with the idea of medical care, touchless sensors are used to monitor patients whose traumatic injury or condition prevents physical contact (due to pain or risk of infection). How about a Wi-Fi-based system that lets them express their needs to caregivers outside the room with a simple wave or a blink of the eye?

What other applications can Wi-Fi-based sensing technology improve or enable? What additional technical barriers exist? I’m eager to hear your thoughts in the Comments section below.


_________________

 2016-09-08 01:07個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 靠汽車自主學習!特斯拉自駕功能大躍進 雷達防撞夯

MoneyDJ新聞 2016-09-12

美國豪華電動車廠特斯拉(Tesla Motors)的自動駕駛(Autopilot)模式備受爭議,5月佛州一起死亡車禍更引起美國國道交通安全管理局(NHTSA)的重視,為此執行長Elon Musk努力了數個月,終於在週日(9月11日)宣布軟體更新,改變Autopilot對雷達、相機和影像處理系統的操作模式,宣稱如此一來類似佛州車禍的意外事故有機會避免,未來一兩週內車主就可自行下載更新程式。

The Verge、紐約時報等外電11日報導,Musk召開記者會表示,特斯拉軟體更新8.0數週內就會釋出,此前Autopilot主要以相機、影像處理系統作為主要感測器,再以雷達確認資訊,但在更新之後,相機和雷達會同步偵測路上的障礙物,使系統提前偵測到過去可能會被忽略的撞擊事故。

Musk說,以雷達作為主要感測器是相當困難的技術,除非能把其他車輛連結至雲端,否則無法達成。他將這項技術稱為「車隊學習」(fleet learning),特斯拉汽車把雷達偵測到的周遭資訊回傳至雲端伺服器,讓所有特斯拉的車輛都能同步學習,辨識路上的障礙物,等於是把每一個用戶當成特斯拉汽車的訓練師。

這麼一來,該公司就能以GPS在地圖上編碼,標示不會威脅到車輛的各種物件(例如天橋等等),以防系統在沒必要時頻頻剎車,Autopilot也能連年進化,識別出正在過馬路的卡車、廢棄金屬、甚至幽浮,並知道應該避開這些障礙物。即使客戶的車款較舊,Autopilot也能持續改善,硬體完全沒有升級的必要。

不僅如此,特斯拉的最新技術還能把雷達訊號彈射到前方車輛的底部地面,偵測被前方汽車擋住之處,是否有何突發狀況,有了這項科技,即使前方車輛在濃霧中事故,跟在後方的特斯拉汽車還是能順利躲開意外。

假如特斯拉偵測到車輛即將遭到撞擊,那麼自動駕駛系統會強制啟動、為駕駛踩下剎車,並轉動方向盤降低撞擊的嚴重度。

另外,駕駛人在開啟Autopilot後,經常會放鬆警惕,導致車禍意外。為此,新版Autopilot將增加警示頻率,若一小時內駕駛人因為手未放在方向盤上而三次受到警告,那麼Autopilot就會自動關閉,駕駛人須停車才能重新啟動。

不過,若當前的交通停停走走,行車時速低於8英里,那麼駕駛人就幾乎完全不需把手放在方向盤上,系統會等到時速上升至45英里後、或是道路開始轉彎、前方有車輛時發出警示。

*編者按:本文僅供參考之用,並不構成要約、招攬或邀請、誘使、任何不論種類或形式之申述或訂立任何建議及推薦,讀者務請運用個人獨立思考能力,自行作出投資決定,如因相關建議招致損失,概與《精實財經媒體》、編者及作者無涉。


_________________

 2016-09-14 00:14個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 Infineon, Google ATAP team up to bring gesture control to IoT

24 May 2016, Infineon Technologies



Using the “Soli” radar technology, the companies hope to deliver greater innovation across various applications such as home entertainment, mobile devices and the Internet of Things devices.

Infineon Technologies and Google ATAP have joined forces to unveil a couple of prototyped devices completely controlled through gestures. The smartwatch and wireless speaker can recognise gestures that replace switches or buttons using the innovative concept from the two companies.

“Gesture sensing offers a new opportunity to revolutionise the human-machine interface by enabling mobile and fixed devices with a third dimension of interaction,” said Ivan Poupyrev, technical project lead at Google ATAP. “This will fill the existing gap with a convenient alternative for touch- and voice controlled interaction.” Andreas Urschitz, president of the division power management & multimarket at Infineon, said: “Since mankind started using tools 2.4 million years ago, this is the first time in history that tools adapt to their users, rather than the other way round.”

Market opportunity

Infineon and Google ATAP aim at addressing numerous markets with “Soli” radar technology. Among these are home entertainment, mobile devices and the Internet of Things (IoT). Radar chips from Infineon as well as Google ATAP’s software and interaction concepts form the basis. Both companies are preparing for the joint commercialisation of the “Soli” technology. “Sophisticated haptic algorithms combined with highly integrated and miniaturised radar chips can foster a huge variety of applications,” said Urschitz.


In addition of their efforts in the audio and smartwatch markets, the developers’ ambitions are more comprehensive: “It is our target to create a new market standard with compelling performance and new user experience, creating a core technology for enablement of augmented reality and IoT,” said Urschitz. While virtual reality technologies could already visualise new realities in the past, users could not interact with these so far. The 60GHz radar application developed by Google and Infineon bridges the gap, as a key technology enabling augmented reality.


_________________

 2017-02-28 23:50個人資料傳送 Email 給 jason
jason
Webmaster



註冊日: 2003-02-15
發表數: 1190


 Sensor fusion: radar and MEMS microphones with audio processor for unmatched voice-recognition

Munich, Germany – February 28, 2017 – Infineon Technologies AG (FSE: IFX / OTCQX: IFNNY) and XMOS Ltd. partner to deliver a new building block for voice recognition. It features a combination of radar and silicon microphone sensors from Infineon and audio processor from XMOS. The devices provide far field voice capture by audio beamforming combined with radar target presence detection. Together, they guarantee for optimal sound recognition and flawless execution for digital voice assistance across a broad range of voice-controlled devices. Target applications are especially smart home as well as smart TV and set top box, secure keyless entry systems, and other voice-operated consumer devices.

The performance of current MEMS microphone limits the effectiveness of systems: when multiple persons are speaking, the veracity and location of the voice source cannot be precisely identified and separated from inanimate object noise. Infineon’s 60 GHz 2Tx/4Rx radar IC with accompanying antenna and the 70dB SNR microphone helps to overcome these impediments. The microphones are based on the dual backplate MEMS technology from Infineon and are well suited for far field voice capturing and beamforming. In addition, improving the SNR of the microphone will further enhance the performance.

The XMOS audio processor analyses the signal data from an array of Infineon’s digital microphones, adjusting the angular and distance data from each microphone to produce a beam at an angle identified by the radar data. A lit LED indicates presence detection and where sound is taken from.

The combination of an Infineon radar and the XMOS beamformer allows the microphone to target a specific object precisely, even with the object moving and indistinctive noise. The microphone delivers a superior user experience compared to existing voice-controlled systems, with clear understanding of commands and flawless execution that prevents unwanted activations of devices. More information is available at www.infineon.com/audiobeamforming.

Inspiring mobile business – Powering the future the secure way!

At Mobile World Congress 2017 (hall 6, booth 6C41), Infineon will demonstrate semiconductor solutions for secured and energy efficient communication in the connected world. Highlights include intelligent street lamps for a smart traffic infrastructure, security gateways to protect connected devices in the smart home as well as new concepts for facilitating the interaction of humans and machines by using voice and gesture recognition. Further information is available at www.infineon.com/5G.


Radar, mic sensors enable clear voice recognition

2 March 2017, Infineon

Infineon’s 60GHz 2Tx/4Rx radar IC with antenna and the 70dB SNR microphone help locate voice source amidst ambient noise.

Infineon Technologies have combined its radar and silicon microphone sensors with XMOS's audio processors to provide far field voice capture by audio beamforming combined with radar target presence detection. Together, they guarantee sound recognition and execution for digital voice assistance across a range of voice-controlled devices. Target applications are smart home as well as smart TV and set top box, secure keyless entry systems, and other voice-operated consumer devices.

The performance of current MEMS microphone limits the effectiveness of systems: when multiple persons are speaking, the veracity and location of the voice source cannot be precisely identified and separated from inanimate object noise. Infineon’s 60GHz 2Tx/4Rx radar IC with accompanying antenna and the 70dB SNR microphone help to overcome these impediments. The microphones are based on the dual backplate MEMS technology from Infineon, and are well suited for far field voice capturing and beamforming. In addition, improving the SNR of the microphone will further enhance the performance.

The XMOS audio processor analyses the signal data from an array of Infineon’s digital microphones, adjusting the angular and distance data from each microphone to produce a beam at an angle identified by the radar data. A lit LED indicates presence detection and where sound is taken from.

The combination of Infineon radar and the XMOS beamformer allows the microphone to target a specific object precisely, even with the object moving and indistinctive noise. The microphone delivers clear voice-controlled systems, with understanding of commands and flawless execution that prevents unwanted activations of devices.


_________________

 2017-03-03 18:00個人資料傳送 Email 給 jason
到 ( 1 | 2 下頁 )