Chen and Nizan Sagiv had the idea when they were at a Depeche Mode 
concert in Tel Aviv five years ago. "While I was busy looking at the 
show, Nizan was watching the crowds," explained Chen, SCENENET project 
coordinator. "He could not help noticing the huge number of faint lights
 from mobile phone screens. People were taking videos of the show. Nizan
 thought that combining all the videos taken by individuals into a 
synergetic, enhanced and possibly 3D video could be an interesting idea.
 We discussed the concept for many months, but it looked too futuristic,
 risky and complicated."
Israel–Europe collaboration
They went for advice from ISERD, 
the Israel-Europe R&D directorate
 , and contacted Prof. Peter Maass, of the University of Bremen in 
Germany, and Prof. Pierre Vandergheynst, of Ecole Polytechnique Fédérale
 (EPFL) in Lausanne, Switzerland, with whom Chen had worked on an 
earlier 7th Framework Programme project, 
UNLOCX .
The result is the SCENENET project, awarded € 1.33 million by the 
European Commission, and coordinated by Chen's and Nizan's 
Ra'anana-based company, SagivTech, specialists in computer vision and 
parallel computing.
SCENENET, which is funded by the 
Future and Emerging Technologies (FET
 scheme) , just as UNLOCX was, runs until January 2016 and consists of 
four European partners: the University of Bremen, Steinbeis Innovation, 
European Research Services, all in Germany, and Switzerland's EPFL.
The first year of the project has seen the team develop the mobile 
infrastructure for the video feeds, a mechanism for tagging them, and 
their transmission to a cloud server. They’ve also developed basic tools
 for a human-computer interface that will allow users to view the 3D 
video from any vantage point 'in the arena' and edit the film 
themselves. This, they believe, will help create online communities to 
share the content, reliving the concert experience together. With this 
in mind, the partners are to study privacy and intellectual property 
rights issues during the next two years of the project.
"We have at the end of the first year, and sooner than expected, 
built the entire SCENENET pipeline based on current state-of-the-art 
components," said Chen. Indeed, the accelerated computer vision 
algorithms they are creating for mobiles is a pioneering effort 
worldwide and leading chipset manufacturers are following the project 
closely.
SCENENET involves several technological challenges: on-device 
pre-processing that requires immense computer power, efficient 
transmission of the video streams, development of accurate and fast 
methods for registration between the video streams, and the 3D 
reconstruction. All of these tasks have to run at near real-time rates.
"We believe that the various components that make up SCENENET, e.g. 
registration of images and 3D reconstruction, have great potential for 
mobile computing and cloud computing, Thus SCENENET offers a huge 
technological breakthrough – in its whole and also via each of its 
components," she added.
Moving beyond concerts
Myriad possible uses for SCENENET are emerging as the project 
develops. Rights and privacy concerns permitting, the technology might 
also be used to recreate other events in 3D, such as breaking news or 
sports, or in the tourism or surveillance sectors. The partners are also
 looking at shooting static, as well as active, objects from various 
angles, to create instructions that can be sent on to 3D printers. The 
mobile-cloud server model could also be used for a host of other 
applications, say the researchers. But for the moment they are 
concentrating on music fans.
"SCENENET revolves around mobile cameras and 3D vision. The invasion
 of mobile cameras and their continuously improved quality has meant we 
are flooded with images we want to enhance and show off. Many devices 
that 'understand' visual inputs are being developed – Google Glass, for 
instance - where most of this work is based on image processing and 
computer vision. 3D vision is becoming more important for better 
visualization of the world on one hand, and easier analysis of the world
 on the other hand," the coordinator explained.
SCENENET is a good example of collaboration in research between the 
EU and Israel. Some 1 600 Israeli scientists have benefited from the 7th
 Framework programme in over 800 projects funded by the EC to the tune 
of € 634 million, in fields ranging from cutting edge research and ICT 
through nanotechnology to energy and health.
"This is an opportunity to be a part of the thriving European 
scientific and industrial communities, and collaborate with leading 
academy partners and companies in Europe," Chen summarised.
SCENENET received research funding under the European Union's Seventh Framework Programme (FP7).
Link to project on CORDIS:
- 
FP7 on CORDIS
Link to project's website:
- 
SCENENET website
Other links:
- 
European Commission's Digital Agenda website
- 
"Uncertainty principles versus localization properties"
- 
HORIZON 2020 - Future and Emerging Technologies