據(jù)悉谷歌在圖像處理方面有望再獲突破,可對不含地理信息的圖片實現(xiàn)精確地理定位。你拍照,我就知道是哪里。
測試中可能遇到的詞匯和知識:
stride 大步[stra?d]
random 隨機的['r?nd?m]
decipher 解釋[d?'sa?f?]
subtle 微妙的['s?t(?)l]
kidney 腎臟['k?dn?]
閱讀即將開始,建議您計算一下閱讀整篇文章所用時間,并對照我們在文章最后給出的參考值來估算您的閱讀速度。
By Murad Ahmed, European Technology Correspondent
-----------------------------------------------------
Google has unveiled a system that attempts to pinpoint the location of where a photograph was taken by analysing the image, as the internet group continues to experiment with advanced “machine learning” technologies.
Though at its early stages, the Californian company’s system is another example of how Silicon Valley groups are making giant strides in artificial intelligence, using the ability to crunch huge amounts of data and spot patterns to develop capabilities far beyond human brains.
Facebook, IBM and a number of start-ups are working on AI features to power devices such as digital assistants, healthcare apps that can diagnose medical conditions and the software that can power driverless cars.
Google’s latest experiment attempts solve a task that most humans find difficult: looking at a picture at random and trying to work out where it was taken.
Humans are able to make rough guesses on where a shot has been taken based on clues in the picture, such as the type of trees in background and the architectural style of buildings. This task has proven beyond most computer systems.
This week, Tobias Weyand, a computer vision specialist at Google, unveiled a system called PlaNet, that is able to decipher where a photograph has been taken by analysing the pixels it contains.
“We think PlaNet has an advantage over humans because it has seen many more places than any human can ever visit and has learnt subtle cues of different scenes that are even hard for a well-travelled human to distinguish,” Mr Weyand told MIT Technology Review, which first reported the news.
His team divided the world into a grid containing 26,000 squares —— each one representing a specific geographical area.
For every square, the scientists created a database of images derived from the internet that could be identified by their “geolocation” —— the digital signatures that show where many photographs are taken. This database was made up of 126m images.
Using this information, the team would teach a neural network —— a computer system modelled on how layers of neurons in the brain interact —— to place each image to a specific place.
Mr Weyand’s team plugged 2.3m geotagged images from Flickr, the online photo library, to see whether the system could correctly determine their location.
Though this means it is far from perfect, this performance is far better than humans. According to the team’s findings, the “median human localisation error” —— meaning the median distance from where a person guessed the location of a picture, to where it was actually taken —— is 2,320.75km. PlaNet’s median localisation error is 1,131.7km.
The news came as Google’s Deepmind group, its artificial intelligence arm based in the UK, announced its first push into medical technology on Wednesday.
The company has created the healthcare project following the acquisition of Hark, a health tech start-up which has created digital tools for patients.
One of its first ventures is “Streams”, an app that aims to give nurses and doctors timely information about patients to help detect cases of acute kidney injury.
請根據(jù)你所讀到的文章內容,完成以下自測題目:
1. What phase is the system in?
a. also in the plan
b. middle stage
c. mature stage
d. early stage
2. Which one is not mentioned as the AI feature?
a. VR glasses
b. healthcare apps
c. software about driverless cars
d. digital assistants
3. How to decipher where a photograph has been taken by PlaNet?
a. mobile positioning system
b. connecting a social network
c. analysing the pixels
d. analyzing user’s historical records
4. Which area is Google’s Deepmind group working for?
a. medical technology
b. information system
c. Cloud Computing
d. doctor-patient relationship
[1] 答案 d. early stage
解釋:這一圖像處理的軟件還處于初級階段。
[2] 答案 a. VR glasses
解釋:文章并未提到虛擬現(xiàn)實眼鏡是人工智能。
[3] 答案 c. analysing the pixels
解釋:PlaNet通過對圖片進行像素級分析的基礎上,與圖片庫中的存儲數(shù)據(jù)進行像素比對,以實現(xiàn)二者之間的最佳匹配。
[4] 答案 a. medical technology
解釋:谷歌旗下的人工智能子公司DeepMind Health已經開發(fā)出一款名為”Streams”的軟件,讓臨床醫(yī)生能夠更快的觀察到醫(yī)療結果。