PDF - Geospatial World
Transcription
PDF - Geospatial World
G E O S P A T I A L I N D U S T R Y M A G A Z I N E GEOSPATIAL WORLD JANUARY 2016 » VOLUME 06» ISSUE 6 | ISSN 2277–3134 www.geospatialworld.net INDUSTRY ANNUAL EDITION 2016 The geospatial industry is going through a great churning to discover new processes and business models. Catch up with the pathbreaking innovations and technologies R.N.I No - UPENG/2010/34153; Registration no: UP/GBD-136/2014-16 Publication: 10th of every month I Posting: 15th / 20th of every month Price: INR 150 / US$ 15 Subscriber’s copy. Not for Sale Y O U R Transforming CiTy resilienCe Through images and daTa analyTiCs To accomplish resiliency, governments require properly planned infrastructure and programs to quickly assess impacts caused by major events. With the mobile imaging and remote sensing technologies of Trimble® Unmanned Aircraft Systems (UAS), governments can quickly accomplish change or damage assessments to prioritize planning and Trimble UX5 Unmanned Aircraft System Trimble eCognition Image Analysis Software Leading image acquisition quality and data accuracy Automate feature extraction to map existing conditions response efforts from macro to micro scale. The Trimble UX5 UAS provides rapid images for mapping and analysis. Coupled with Trimble eCognition® software, city staff can import, fuse, interpret, and analyze data from the UX5 as well as other sources to achieve a greater understanding of their environments. Combining powerful images and analytics improves accuracy in taxation, aids effective transportation and infrastructure design, and promotes healthier environments more in balance with nature— all resulting in greater city resilience. Discover how Trimble technologies support the resilience of your city at www.trimble.com/RC-Summit © 2015, Trimble Navigation Limited. All rights reserved. Trimble, the Globe & Triangle logo, and eCognition are trademarks of Trimble Navigation Limited, registered in the United States and in other countries. All other trademarks are the property of their respective owners. GEO-067 (10/15) Comprehensive geospatial deliverables for resiliency planning and monitoring Trimble is a 100 Resilient Cities Platform Partner INTRODUCING HEXAGON SMART M.APP® Hexagon Smart M.Apps are streamlined and lightweight market applications that solve unique business problems. Each Hexagon Smart M.App combines multisource content, sophisticated analytics, and tailored workflows, fused together into a dynamic user experience. SMART CITIES, SMART NATION BUILDING SMART CITIES THROUGH GEOSPATIAL INTELLIGENCE • Create a common operational picture for both spatial and non-spatial data. • Enhance livability, workability and sustainability by sharing data. • Disseminate real-time data to desktop, web, and mobile devices for asset and infrastructure management, transportation, utilities, telecom, public safety, and security. CONTACT US in-sales@hexagongeospatial.com LEARN MORE hexagongeospatial.com CONTENT 7 VOLUME : 6 ISSUE : 6 Editorial Tech Trends Industry Outlook The changing dynamics of the geospatial industry is changing the way we use technology. As the focus shifts from hardware to software, and standalone products give way to system integration and solutions, businesses are getting impacted, ushering in an era of consolidation, collaborations and partnerships. There’s revolution in data acquisition, Big Data is becoming bigger, apps are being more popular and things are getting more and more connected in the era of IoT. Catch up with some of the most prominent technology trends and how they are changing the industry. 36 Small Satellites 10 Creative destruction of GIS is upon us. Are we ready? Small sats have come into their own in big boys’ realm 40 UAVs 12 Ola Rollén Hexagon 14 Bryn Fosburgh Trimble Drones take wings, but still bound by regulations 27 Wade Larson UrtheCast 28 Robert Zitz Leidos 16 Bhupinder Singh Bentley Systems 29 Juergen Dold Hexagon Geosystems 48 Mobile Mapping 18 Ray O’Connor Topcon Positioning Systems 30 B.V.R. Mohan Reddy Cyient 52 Apps 20 Amar Hanspal Autodesk 31 Manish Choudhary Pitney Bowes 44 Big Data Future is in analytics & interconnected Big Data Proliferation in sensors driving mobile mapping Apps the new way to go, Thanks to mobile devices 56 Automated Cars Cars will no longer be owned, but called up when required 22 24 K.K. Singh Rolta Dr Kumar Navulur DigitalGlobe 32 34 Michel Stanier Teledyne Optech Amit Somani ADCC Infocad 60 Smart Cities Citizen-friendly solutions the key, not technology fixes 64 BIM + Geospatial 26 Stuart Blundell Harris Geospatial Solutions 35 Anup Jindal RMSI BIM + Geospatial = Environmentally sustainable cities & infrastructure 5 • Geospatial World • January 2016 Policy Pulse Policies guide the total strategic effort of the geospatial industry. The biggest push on this front has been in the shape of open data initiatives taken by various governments. After all, what is the point of having great content if you are not going to liberate it? 68 UN-GGIM: The Way Forward 72 Barbara Ryan Group on Earth Observations 77 Kshemendra Paul Information Sharing Environment, US 74 Scott Pace Space Policy Institute, George Washington University 75 Sue Gordon NGA, US 76 Tim Trainor US Census Bureau 78 Nigel Clifford Ordnance Survey, UK 79 Jeffrey Harris USGIF, US 80 A.S. Kiran Kumar ISRO, India BIM 360, Autodesk ContextCapture, Bentley Product Watch Get updated on some of the most promising products and services from top companies Pages 82-90 Falcon 8, Topcon VZ-400i, RIEGL UltraCam v5, Microsoft AppStudio, Esri Intergraph NetWorks, Hexagon Focus3D X 30, Faro LuciadRIA, Luciad Galaxy, Teledyne Optech PageWide XL, HP East View Nautiz X8, Algiz RT7, Algiz 10X, Handheld TerraGo Edge, TerraGo ConflateX, RMSI Pix4Dmapper, Pix4D Summit UAS, DAT/EM Systems 6 • Geospatial World • January 2016 From the editor Prof Arup Dasgupta Managing Editor, arup@geospatialmedia.net 2016 could be a year of ‘democratization’ E very time when the month of January comes around, I begin to wonder what earth-shattering geospatial events am I going to predict for the New Year in the January edition. This year, I was intrigued and then convinced by the arbitrariness of this concept of New Year after someone on Facebook had recently concluded that the concept is fallacious as it just represents the completion of one more orbit of the earth around the sun from an arbitrary starting point. It is not even a heliocentric reference point like the winter solstice; summer as those ‘down under’ will aver, which itself is an arbitrary construct. Going through various articles slated for the January 2016 edition, it dawned on me that the world in general and geospatial world in particular does not race from event to event timed by this arbitrary epoch. Rather it is a process of continuous change, of waves of evolution as it were. At intervals some technology arrives to disrupt the status quo but soon gets absorbed in the scheme of things through a process of adoption and adaptation. For example, can we mark a point in time when say, remote sensing became active? Rather, do we not see a blip that was aerial surveys, then handheld photography by astronauts, then TIROS and then several blips like ERTS, RESURS, SPOT and IRS that seem to blend into a continuum of satellites with more and more advanced features. Therefore, instead of trying to be a Geospatial Nostradamus, I shall limit myself to a few observations based on the current situation of the geospatial world around us. My first impression is that the leitmotif for 2016 seems to be ‘democratization’. This is reflected in the shift towards consumption-oriented business models. Consumers are looking for localized solutions, but are not interested in putting together solutions themselves from scratch, buying data, systems and software, coding their own applications and managing workflows. In this consumer-centric holistic approach, the data acquisition is purposed and the workflow is integrated into the hardware and software such that the consumer is able to get solutions on demand and at their location which could be at their desk or on site. This indicates that geospatial services will continue to grow in the form of integrated applications, or apps. Apps bring the real world to the digital world and thus helps the user to understand the problem and apply the right solutions to it. Chairman M P Narayanan Publisher Sanjay Kumar Managing Editor Prof. Arup Dasgupta Editor — Defence & Internal Security Lt Gen (Dr) AKS Chandele (Retd) Editor — Mining (Hon) Dr. Hrishikesh Samant Editorial Team Bhanu Rekha Anusuya Datta Ishveena Singh Amit Raj Singh Design Debjyoti Mukherjee Product & Marketing Team Harsha Vardhan Madiraju Sanskriti Shukla Vijay Kumar Singh Disclaimer Geospatial World does not necessarily subscribe to the views expressed in the publication. All views expressed in this issue are those of the contributors. Geospatial World is not responsible for any loss to anyone due to the information provided. Owner, Publisher & Printer Sanjay Kumar Printed at M. P. Printers B - 220, Phase-II, Noida - 201 301, Gautam Budh Nagar (UP) India Publication Address A - 92, Sector - 52, Gautam Budh Nagar, Noida - 201 301 India. The edition contains 92 pages including cover Geospatial World Geospatial Media and Communications Pvt. Ltd. (formerly GIS Development Pvt. Ltd.) A - 145, Sector - 63, Noida, India Tel + 91-120-4612500 Fax +91-120-4612555 / 666 Price: INR 150/US$ 15 7 • Geospatial World • January 2016 International The old cliche ‘the consumer is always right’ seems to be the indication for geospatial activities in 2016 Apps will have to be platformagnostic, based on standards and deployed on the Cloud. Apps will cover a very wide range from 3D solutions for engineers to location-based consumer solutions for tourists. Apps will need data from various sources; geospatial data from satellites, UAVs and ground instrumentation which could be human operated DGPS or automated purposed sensors, like meteorological instruments. Considerable data will also come from other data streams, like GPS data from smartphones, financial data from electronic financial transactions and from RFID sensors mounted on mobile assets. Another major source will be unstructured data coming from crowd sourcing and from social networks. The processing of large data streams, both structured and unstructured, will call for Big Data analytics. While this is already well established in business intelligence applications, it is yet to make significant inroads in the geospatial world. However, with the range of data collecting systems increasing from large and small satellites, aircraft and UAVs, static and mobile sensor networks and crowd sourcing, it is only a matter of time before Big Data analytics becomes a part of the armoury of the geospatial professional. The publication and discovery of these data sets and analytic solutions will depend strongly on the Cloud. Apps development is open to independent developers for consumer solutions. This trend is extending to professional solutions as well. As solutions become complex, their realization will require partnerships between government institutions and private players, between data generators, hardware and software providers and developers. Partnerships will require sharing of data and intellectual property and this will have to happen in an environment of security which encourages confidence without become overbearing. Univerzalisation of apps will need certain policy initiatives like standardization and open data. While standardization of data has been achieved the open access to data is still work in progress. Data for public good is easily available but other data which has economic implications still requires to be purchased. Government institutions in many countries are still grappling with this issue. Another issue is that of privatization of solution providers. Clearly, where public good is involved, like at times of disaster, the government has to take a lead role. However, in areas like inventory, infrastructure, business and consumer affairs the private sector will play a bigger role. The old cliche “the consumer is always right” seems to be the indication for geospatial activities in 2016. On that thought, let me sign off with best wishes for the next 366 rotations of the planet Earth about its axis. Remember, 2016 is a leap year! Advisory Board Ahmad Fauzi Bin Nordin Sr Director General of Survey and Mapping (JUPEM), Malaysia Aida Opoku-Mensah Special Advisor: SDGs/Agenda 2030, UN Economic Commission for Africa Barbara Ryan Secretariat Director, Group on Earth Observations Christopher W Gibson Vice President & Executive Committee Member, Trimble Derek Clarke Chief Director-Survey and Mapping & National Geospatial Information, Dept of Rural Development & Land Reform, South Africa Dorine Burmanje Chair-Executive Board, Cadastre, Land Registry and Mapping Agency (Kadaster), The Netherlands Ed Parsons Geospatial Technologist, Google Greg Bentley CEO, Bentley Systems Jay Freeland Retired President & CEO, FARO Prof. Josef Strobl Chair, Department of Geoinformatics, University of Salzburg, Austria Kamal K Singh Chairman and CEO, Rolta Group Kumar Navulur Director, Next Generation Products, DigitalGlobe Mark Reichardt President and CEO, Open Geospatial Consortium Mladen Stojic President, Hexagon Geospatial Mohd Al Rajhi Asst Deputy Minister for Land & Surveying, Ministry of Municipal & Rural Affairs, Saudi Arabia Sandeep Singhal Director, Cloud Storage, Google Stephen Lawler Senior Vice President, Worldwide Engineering, Cyanogen 8 • Geospatial World • January 2016 Vanessa Lawrence Inaugural Co-Chair, UN-GGIM NEW RIEGL VZ-400i High Performance 3D Laser Scanner » Ultra High Speed Data Acquisition » Survey-Grade Accuracy » Extremely Robust & Reliable » Real-Time Registration & Processing » Cloud Connectivity via Wi-Fi and 4G LTE Farther, Faster, Better: The NEW RIEGL VZ-400i is Redefining Productivity! This evolution of laser scan engine technology is based on its new innovative processing architecture. With advanced processing technology, data acquisition and simultaneous geo-referencing, filtering and analysis in real-time. The New RIEGL VZ-400i is the fastest end-to-end Terrestrial Laser Scanning System on the market, setting the benchmark in 3D Laser Scanning, again! Ultra High Speed Data Acquisition with 1.2 MHz laser pulse repetition rate | 1 m – 800 m range | 5 mm survey grade accuracy | real-time registration & processing | Cloud Connectivity via Wi-Fi and 4G LTE user friendly touchscreen interface | MEMS IMU for pose estimation | advanced flexibility through support for external peripherals and accessories | high end camera option Stay connected with RIEGL www.riegl.com RIEGL LMS GmbH, Austria | RIEGL USA Inc. | RIEGL Japan Ltd. | RIEGL China Ltd. Industry Outlook Creative destruction of GIS is upon us User engagement is driving the industry as mainstreaming of geospatial data and technology is unleashing its real value and potential I n a futuristic article in the annual edition (January 2010) of Geospatial World, Ed Parsons, Geospatial Technologist at Google, had predicted that the ‘Creative Destruction of GIS’ will soon be upon us owing to the broad trends in technology, Internet and the increasingly ubiquitous use of geospatial data and applications. Today, GIS and geospatial technology have reached that existential moment where the belief in being ‘exclusive’ is fast getting replaced with the realization that geospatial is a given, integrated and invisible capability. This mainstreaming of geospatial data and technology has unleashed the value and potential of this powerful technology in umpteen ways, silencing the critics who often said geospatialists were enamored with the exclusivity of the technology and the skill it takes to develop that exclusive capability rather than the benefit it creates. Today, the user is the centre of geospatial 10 • Geospatial World • January 2016 Are we ready? ecosystem and every industry trend can best be seen as a result of the deepening dialogue, clear understanding of the needs and increasing trust and engagement between technology providers and technology users. Consumerization of professional space Companies are increasingly leveraging Cloud-based technologies to deliver value from their labs to customers’ job sites/offices. The ability to deliver Cloud services through existing applications and being able to deliver updates is another reason innovation cycle is rapidly accelerating. The agility and flexibility of the Cloud in catering to and engaging with the user in a two-way communication is also driving software business, a sign of maturity of the industry. The software subscription models are perfectly aligning with customer project cycles, allowing them to select what they need and when they need. Smarter hardware Quick change is the norm in a competitive business environment making companies pioneer and/or reinvent themselves, sometimes by design and sometimes because of necessity. Companies are increasingly aware that their bottomlines are determined by the value their technology brings to customers. So, they primarily focus on providing an end-to-end workflow solutions. The trend has also increased the use of open source software and open data along with proprietary software allowing hybrid solutions to proliferate. Collaboration and connectivity across the workflow is improving productivity, enhancing transparency and quality. However, no single company has all the requisite technology components of the solution value chain and companies are increasingly adding cost-effective and competitive offerings either by acquiring, partnering or collaborating with others, thus creating a rich culture of innovation and professionalism within the industry. Shared information infrastructure No solution is competent enough to deliver the requisite outcomes of a user without appropriate data. New and sophisticated sensor technology is causing an exponential increase in the volume, type, speed and ease of access to data. This data revolution and improved analytics are creating unprecedented possibilities for geospatial industry facilitating rapid turnaround for informed intelligence and decision making. Access to the latest data is the key to unlocking this intelligence and so conventional hardware/software companies are increasingly adding data capabilities to their inventory. They are also partnering with other data providers and sharing data and information repositories. Hexagon Geospatial’s recent partnerships with Ordnance Survey UK, Airbus Defence and Space, and BlackSky are examples of shared information infrastructure trend. The increasingly positive senIntuitive timent around open data is allowing huge software repositories to be shared as stakeholders become increasingly aware of the returns of open data to businesses and economy. The versatility of Cloud and the need to solve increasingly complex problems have vastly evolved the way data and information is shared. The ability to manipulate and process data in a shared high-performance End-user computing environment is allowing a engagement user to access data from any location. The model economics of moving data, processing and hosting services by commercial Cloud providers also has become an accepted norm. The key drivers behind the concept of “sharing” are open source Cloud platforms as a service (PaaS) providers such as Cloud Foundry and commercial Seamless Cloud services from biggies workflow like Amazon, Google, IBM and Microsoft. Some believe that soon, shared information will be the expectation, and, in the eyes of many, a social responsibility. Integration is the key In the end-user engagement model, hardware is becoming smarter, software is becoming intuitive, workflow is becoming increasingly seamless and solutions have more ease of use. However, to truly benefit from systems integration, virtually every aspect of the workflow must be considered. Geospatial technology must be integrated in such a way that when it is employed on any aspect of a project, that data — that solution — is preemptively, or actively available, to solve customers’ current and future needs. The bottom line, in the entire scheme of things, is value creation and building trust. Easy-to-use solutions Let's see what our experts are Saying Industry Outlook We can’t Solve Today’s Problems with Yesterday’s Technology Many organizations that have not been traditional geospatial customers are starting to see the inherent value in location information and are seeking ways to acquire, manage, disseminate and utilize this content 12 • Geospatial World • January 2016 O ne of the trends in the industry today is the shortening product/ tech innovation lifecycle — not only in terms of development, but also in regard to the entire decision-making process. Individuals and organizations see value in geospatial information, but not necessarily through the traditional depiction in a static map. The expectation today is for a simple, understandable answer that provides unique insight into complex problems. This is a trend that we are seeing in many markets and has been an impetus for us to innovate the way we deliver our technology, providing a pragmatic approach to building solutions quickly and in ways that make sense. We are definitely shortening the decision-making process through the Hexagon Smart M.App — combining fresh content, targeted workflows and meaningful analytics to provide insights in a simple way. We are providing online access to near-real-time data through partnerships with companies like Airbus Defence and Space and Ordnance Survey UK, and fusing it with best-in-breed geospatial tools for customization. We have created an ecosystem of empowering partners with specific domain knowledge to build not just a map, but an information experience based on streams of data and powerful, 360-degree analytics. End-to-end workflow with solutions Hexagon doesn’t look at the world of hardware or software or apps or content separately. Our primary focus is to provide an end-to-end workflow with solutions that make sense. And in many cases, this requires hardware, software and content, together with industry-built solutions that actually solve problems. However, we do continue to see numerous opportunities in our hardware, especially in the geospatial space. Technologies like mobile mapping, terrestrial LiDAR mapping and UAS’ are growing, and we expect to see this trend continue. One of the exciting things about the geospatial field right now is that many organizations that have not been traditional geospatial customers are starting to see the inherent value in location information and are seeking ways to acquire, manage, disseminate and utilize this content. As more of these customers come on board, they will need a way to get the information they require. More out-of-the-box thinking required At the end of the day, if you do not understand a user’s specific workflow or problem, don’t expect them to use your solution (even if it is inexpensive). As a whole, I think we haven’t been thinking outside the box. Innovations from 20 years ago are providing a baseline for the technology we are now offering to a new, modern user base. And yes, when you try to extend ‘as is’ technology from 20 years ago to a new audience, you can expect the uptake to be minimal. So we cannot solve the problems of today with technology from yesterday. For this reason, we have to innovate ourselves to increase the uptake of geospatial technology to a broader and much wider audience. What customers really want is an answer to a business question. Everything else is just a tool to derive that answer. For this reason, the Hexagon Smart M.App is not about delivering products or technology, but rather about providing access to the answers an individual seeks. For example, Geosystems GmbH, a Hexagon Geospatial partner in Germany, has built a Cloud-based application that can take the data collected by UAS platforms and process it on the Cloud, generating image mosaics and 3D point clouds from the raw data. This enables a UAS company to collect more data instead of being tied to their computer processing data. Data should be interoperable, easily working in as many platforms as is feasible. Hexagon has always valued this methodology, connecting to and partnering with as many organizations as possible. Access to the latest data, whatever the format, empowers you to make the smartest decisions. You can see this commitment through Hexagon Geosystems’ and Hexagon Safety & Infrastructure’s partnerships with other software companies and system integrators, as well as Hexagon Geospatial’s recent partnerships with Ordnance Survey UK, Airbus Defence and Space, and BlackSky, among others. many people as possible. Last, in order meet industry needs, we must form vertical and regional partnerships to make our solutions available on a local and global basis. Mainstream IT companies like Amazon, Microsoft, Apple and Google are adopting location capabilities by buying-out geospatial players. If anything, this provides us much greater opportunities in the marketplace. I think it is actually a sign that the market is coming to realize the vital importance of location-based data. The explosion of geospatial data is outpacing what even the experts predicted. Everything has a geospatial component, especially as more and more everyday objects become geospatially enabled. Lamp posts will soon tell us not just that their bulb burned out, but exactly where they are located, allowing us to quickly schedule maintenance trips. Every purchase is made at a specific location, allowing us to not just understand where things are purchased, but to better plan the logistics to get the right products to the right places at the right time. This can completely transform the way we do business, and Hexagon can provide the technology to power this understanding and help our customers shape smarter change. Partnerships add to strengths Hexagon greatly values the strength that can come from partnerships. There is such a deluge of data being collected about our dynamically changing world, and any single piece of that data could provide the key to unlocking major innovations in how we understand the earth. As I have said, we have been actively building partnerships with data providers to help ensure that our tools work as seamlessly as possible. Similarly, we are partnering with other platform providers to ensure our content gets into the hands of as Ola Rollén President and CEO, Hexagon Geospatial World • January 2016 • 13 Industry Outlook Consumer Market Driving the Professional World From providing horizontal services, geospatial is moving to providing solutions to vertical markets, which is leading to the growth of overall geospatial industry 14 • Geospatial World • January 2016 T he consumer world expects a new technology gadget every 12 to 18 months. This expectation has now been driven into the professional marketplace and applies to both hardware and software. In the past, technology or products from the professional world were migrated to the consumer world. Today, that trend has shifted and it is the consumer world driving and seeding the development of the professional world. All companies have a ‘not invented here’ component in their culture. This is good because it creates a balance of buy versus build. In addition, it helps drive innovation and maintain an effective cost and competitive offering. However, Trimble is looking to utilize its industry domain expertise to develop applications leveraging both consumer and professional platforms. Today, if you look at the mixed reality environment, either with a Google Tango or Microsoft HoloLens, these technologies were developed for the consumer world. This will provide the customer an innovative platform with a much quicker time to market, as compared to if the technology was developed internally. Solution takes the cake People ask if we are a hardware or software company. My answer is that we transform the way the world works by providing solutions that improve the productivity, quality, and safety of a particular industry. This is accomplished by utilizing technology (hardware and software) that integrates and optimizes an industry workflow continuum. We have broadened our portfolio to include more software, but this does not diminish our hardware offering. The comment that software and apps ‘take the cake’ is probably not relevant. It is the solution that ‘takes the cake’. Think of your smartphone. If it was only hardware and no apps, would it be anything more than a phone? If you had the apps but no smartphone platform, would they improve productivity? In the last decade, geospatial companies have primarily been concentrating on professional services — either surveying, GIS or photogrammetry that have been providing deliverables to specific industries, like construction — whether it is heavy civil construction be it roads, railways, airports or building construction or mining or specific industries. From providing horizontal services, geospatial is moving to providing solutions to vertical markets. If you take the case of mining industry, one needs to plan-design-mine the ore-reclaim. Connectivity across this workflow and collaboration enhances the productivity. If I can collaborate across the stakeholders, I can get visibility, I can improve transparency that will really enhance quality and productivity. So, it is the customer and the customer’s workflow that is driving the integration. And we as manufacturers have to look at making our solutions more connected. Consolidation of the industry has been part of this process. As you expand geospatial to touch more vertical markets, that grows geospatial industry as a total market as well. All of us have a set of solutions we offer to a particular industry. A customer needs to be able to integrate those solutions. Cooperation is needed because a customer may have different brands for different solutions because they believe they are the best-in-class solutions. As industry providers, we must provide integration across that workflow or continuum. Hence, our ecosystems, whether from Trimble’s perspective (Trimble Connect) or from others who offer similar collaboration services, we must have interoperability in-built into the systems. OGC drives geospatial standards and on the building side; BuildingSmart has driven standards to allow interoperatbility. As manufacturers, we need to facilitate tightest interoperability by sharing APIs and providing that connectivity. Localized, not generic solutions I think geospatial uptake is highly penetrated in the consumer world. Smartphones have GNSS, a mapping function, and access to location based services. Several non-traditional markets/players are increasingly adopting geospatial technology and that offers more opportunities for those who are geospatial professionals and for those of us who offer solutions to geospatial professionals and people who need geospatial data. For the professional world, it depends on the solution type and geography. Solutions like the traditional mechanical total stations are highly penetrated worldwide. Solutions like imaging solutions have low penetration throughout the world. The challenge for all geospatial Bryn Fosburgh Senior Vice President, Trimble Navigation providers is to have their solutions localized for a particular region and industry. A generic solution usually only meets 80% of the customers requirements. Deep industry localization along with the corresponding market localization, coupled with localized service and support through the combination of local distribution networks and manufacturer offices, enables increased market penetration. It’s not that location has suddenly become important. What has happened is you are finding technologies that come from consumer world driving themselves into the professional world. In the past, the professional world spawned technologies that were lapped up by the consumer world. But now, the trend has reversed. Take Google Tango, which is a mixed reality product, for example. Or even Microsoft HoloLens. In our case, we are looking at those technologies which we call/belong to mixed reality that deal in the world from reality to virtual reality. These are mixed reality platforms that would be developed for gaming industry or other type of applications. We are now looking at how can these be applied to cadaster world, mapping or heavy civil construction. This is really an interesting change. Industry Outlook The Advent of Consumption Oriented Business Models T he geospatial industry is becoming a more engaged industry, and it is Google who changed the dynamics for us. A company whose revenue is driven by advertising has made the map an everyday commodity. This has upped the bar of people’s expectations. If a map can be accessed through a simple interface, why is the wealth of everything else not being explored in a similar manner? What is more, the innovation cycle is rapidly accelerating across the world. The consumerization of hardware, like drones and cheap sensors, combined with the Cloud, has served as an inflection point for the industry as a whole. Companies have been forced to come up with new consumption-oriented business models, as opposed to the business models of ‘pay in advance’. You can track exactly what is being consumed, which is much more efficient. People don’t mind paying for what they have consumed. Rapid innovations are also leading to collaborations in the industry, which is good for the industry as a whole. If companies work with like-minded firms to exchange software and Bhupinder Singh Senior Vice President, Bentley Systems 16 • Geospatial World • January 2016 service-orientation, users can access both the services from a common interface. If both companies are providing online services, why not provide them on one common portal? Ease of use is essential Today, no software vendor can get away with a product that is not easy to use. The new generation of professionals entering the workforce today has grown up on tools which they have used and learnt without any training. When I joined the workforce 30 years ago, I was made to attend four training classes before I was allowed to step into my department. My on-boarding was one month of training. Today, we have only 10-15 minutes of learning modules. The conditioning of the workforce is such that they are demanding results faster than people realize. Companies need to understand that software must be intuitive, so people can learn it without much training. Users today expect seamlessness. They want to start the job on one device, continue on the second device, and then move on to the third. We are growing up in a world of digital natives. The next generations of kids are very comfortable in the digital world, and they demand software that reflects that ease of their capabilities. We should be able to meet that demand and deliver software that would appeal to them just like a Hollywood movie would. Companies need to solve workflow challenges every day. And these are hard challenges because you have products that have been around for a long time. Over time, as you add more and more capabilities to these products, without removing the obsolete capabilities and simplifying them, the products become quite complex. Shared information and security Infrastructure providers are reaching that stage where the storage and compute costs are decreasing rapidly, while We should be able to meet the demand of the next generations and deliver software that would appeal to them just like a Hollywood movie would the amount of data we are capturing is growing. The dimensions of the cost are being replaced by the dimensions of security. We are at a stage where we need to trade off between accessibility and security. And that trade-off is something that the industry is still grappling with. How do you figure out the right level of secure access, especially with critical infrastructure like national security infrastructure? From the standpoint of the industry, this is something we need to work on. Digital rights managing capability is being built into platforms, but we need to be able to surface them in a way which would allow governments to be more comfortable while using the Cloud. It is more likely that private Clouds will evolve in response to these kinds of trends. We want to offer a portal through which other people’s data is accessible. By leveraging Web map services through conceptual design tools, our Cloud services can navigate multiple data providers, and the end users can navigate and engage in a commercial model with the data providers. Our power lies in being able to federate across other companies’ data sources, and there are many of them. We want to make sure our users can find them through our interfaces and then leverage them depending upon what their needs are, and at what scale. Service-orientation to adopt and absorb Service-orientation is a trend that has continued. Our mission continues to being able to define the services, make the services discoverable, being able to federate amongst them and aggregate them. We, at Bentley, have a broad set of services to offer, and with every new capability that we add, we think what service-orientation we can offer with that. Sometimes, we have to wait for a certain hardware to be available before we do that. At other times, we are dependent on some other tool for it to become useful. But, at the end of the day, whenever we come up with anything new, we think of what sort of service orientation will it have to deliver its value, because that is the most flexible way for users to adopt and absorb. That said, there is still a lot of work to be done to get the data sources where they are locked up into the user interfaces. There are several places where a lot of value and lot of decision support has to happen, analysis needs to take place, and careful orchestration of data has to happen. Geospatial World • January 2016 • 17 Industry Outlook We are Entering a New Era of Expectations T Companies are increasingly taking to a systems integration approach as it is a strong driver for business growth, but, that is not to say that the demand for specific application-based products has declined 18 • Geospatial World • January 2016 here is no doubt that the speed of communications along with rapid prototyping, thanks to 3D printing and other “maker” technologies, have dramatically accelerated our vetting processes for new ideas, and compressed our time-to-market for new products. Topcon lives by the mantra of TM-1 — to the market first! We are entering a new era of expectations that are shattering business models in many industries. If any company in our industry thinks it can meet these expectations alone — which can change faster than most businesses can effectively adapt — I believe it will quickly suffer declining sales and market share. Furthermore, they become a target for absorption, probably by one of the companies they should have been partnering with all along. Open architecture for information availability The new generation of geospatial ‘consumers’ will expect an open architecture in regard to information availability. Companies that believe they gain some type of marketing advantage by trying to retain exclusive systems will ultimately find themselves at a great disadvantage. Shared information will be the expectation, and, in the eyes of many, a social responsibility. Companies are increasingly taking to a systems integration approach as it is a strong driver for business growth, but that is not to say that the demand for specific application-based products has declined. So we are taking the approach that neither can dominate our development model. The beauty for the majority of our end-users — whether fully integrated via our enterprise infrastructure, or a stand-alone application — is that the flow of data can be shared based upon the need of the project. Hardware is as important as software That is why it would be fallacious to say that software and apps have scored a march over hardware. Apple would not be where it is without the iPod, iPhone or iPad. Likewise, Topcon wouldn’t be where we are without our 3D machine control systems and innovative robotic total stations. Of course, just as important, we also wouldn’t be headed where we are without our MAGNET software or Sitelink3D site management solutions. The reality is that both hardware and software are central to our business model. Our plan is simple — we want to make the most innovative and productive hardware and software. Period. From a revenue point of view, it really makes little difference overall as our objective is to sell the complete solution. Workflow adaptation and educating consumers Cost and ease-of-use always play a role in adoption of any product, but I don’t think that either are a barrier for exponential ‘uptake’ in geospatial technology. There are two much more critical factors. The first is the workflow adaptation. To truly benefit from systems integration, virtually every aspect of workflow must be considered. Geospatial Ray O’Connor President and CEO, Topcon Positioning Systems technology must be integrated in such a way that when it is employed to any aspect of a project, that data — the solution — is preemptively, or actively available, to solve future issues. Secondly, and these really go hand-in-hand, we must help educate our customers and their workforce in understanding and applying a solutions approach. They know how to gather extreme amounts of data. The key is adapting their workflow and workforce to benefit from it in the most widely productive manner. A number of mainstream IT companies, like Amazon, Microsoft, Apple, Google and many others are not only integrating location capabilities into their products and services, but they are developing consumer opportunities that pull precision geospatial technology along with it. The advantage for a company like Topcon is that we are far more than a ‘location provider’. The widespread credibility and adoption that these companies bring to our core technologies helps accelerate our efforts to make our markets more receptive to integrating geospatial solutions into their workflow. Industry Outlook The Goal is to Make the Best Decisions Using Up-to-Date Information W e are increasingly leveraging Cloud-based technologies to deliver ongoing value to our customers, and to swiftly drive innovations from our labs out to our customer’s job sites and offices. Given the wide use of geospatial data in every sector and the need for real-time access for decision making, we see Cloud-based technologies and tools drive faster and more informed decisions. From agile software development and investments in startups to new projects, we are driving innovation across our organization and ecosystem. For example, the ability in our tools [like, InfraWorks 360, ReCap, Memento] to remove traditional data silos and create single 3D high-res accurate models and views combining CAD, BIM, LiDAR, sonar, etc., allow us to create a digital mirror of the physical world. It also reflects the speed of this innovation lifecycle. 20 • Geospatial World • January 2016 Software via the Cloud Subscription offers two benefits: firstly, it aligns very well by continuously delivering innovation to our customers; and secondly, a subscription model also aligns well with the project cycles of our customers, allowing them to select what they need, when they need it. We know that customer business cycles change as do the scope, size and type of projects they work on. This allows them to put the maximum set of tools to deliver a particular client’s project effectively and efficiently, and gives them the flexibility to change that toolset for the next project. It’s a win-win scenario in both directions. Complete workflow solutions Customers care about getting their work done, not about individual applications. Service-oriented architectures and better interoperability between products are the hallmark of substantiated modern solutions. That is why Autodesk has been developing solutions, such as, InfraWorks 360, which offer civil infrastructure and geospatial professionals a data agnostic tool for getting their work done. Projects carried out in the architecture, engineering and construction [AEC] industry are fundamentally fragmented by nature, because there are so many diverse disciplines at play across distributed locations. The ability to collaborate across companies, disciplines and locations is game-changing for the AEC industry. It is finally allowing us to put data at the center of a customer’s experience as opposed to applications, and the benefits of this change are profound. Imagine the ability on a project to share rich 3D data in real-time for decision making, allowing project members to view that information on a device that best suits their needs. This greatly reduces timelines for RFI and can even address change orders before they happen. Amar Hanspal Senior Vice President, Products, Autodesk Companies are increasingly taking to a systems integration approach as it is a strong driver for business growth, but that is not to say that the demand for specific application-based products has declined Cost and ease of use The time required to adopt BIM — relative to a learning curve and training time — combined with the need for ease of use for BIM solutions are the key investment-related factors challenging the adoption of BIM within the global geospatial community. With this in mind, we are working to make our BIM software solutions easier to use. And we are building apps better suited to the various persons involved with projects. For example, instead of expecting a supervisor to use a BIM authoring tool, we are providing solutions, such as BIM 360 Field, which is geo-location aware and has the right level of functionality a supervisor in the field wants and needs. We are developing several tools using this approach, and making it easier to use by presenting information within the flow of their work process, as opposed to making users change and learn a way of thinking and working completely different from what they already know. Collaboration and partnerships Nothing gets done by individuals alone. So many projects unfortunately wind up being like orchestras without conductors. We must find a way to get everyone involved with a project on the same page, and tapping each player’s knowledge to achieve the best results. Obviously, easy-to-use, Cloud-enabled collaboration solutions, along with BIM, marks a huge step forward for the global AEC industry. The goal is always going to be to make the best decisions using up-to-date information. Geospatial World • January 2016 • 21 Industry Outlook 80% of Data has a Locational Context Businesses are investing in IT systems to unlock the hidden value of data — not only internal business data, but the vast quantities of social and IoT data to gain deeper insights 22 • Geospatial World • January 2016 T he technology industry continues to remain in a state of flux with constant innovation and advancement. The advent of social, mobile, Cloud, and increasingly the Internet of Things has led to a growing deluge of data, in excess of trillions of bytes being generated daily. It is generally acknowledged that around 80% of data has a locational context. As a result, businesses around the world are realizing that adding a geospatial dimension to data provides insights far deeper than is otherwise possible. Therefore, digital map data is rapidly becoming an integral and essential part of decision support systems. Constant advancements in technology is cutting down gestation periods and shortening innovation life span. Convergence, integration and pervasive Cloud-based business solutions are key trends that are driving and shaping the future growth of geospatial technologies. The convergence of geospatial technologies with industry-specific information systems has enabled the harnessing of true potential of geospatial information for improving the productivity and efficiency of enterprises cutting across industries. In today’s world, geospatial technologies are significantly enhancing the quality of decision support systems resulting in more effective business and public governance processes. Platform-agnostic integrated solutions Our approach has been to develop integrated solutions that are platform-agnostic. Envisaging the market trends over the past few years, we have developed a repository of sophisticated IP that integrates geospatial information with mainstream IT systems. These Cloud-enabled IP-based solutions are at the heart of numerous solutions for effective decision making enriched with locational intelligence. This approach has placed Rolta at the forefront of IT companies that provide field-proven solutions leveraging geospatial technologies to solve real-world problems. The relentless innovation in the last decade has clearly resulted in hardware becoming commoditized. Furthermore, the advent of mobility and pervasive computing enabled by the Cloud has drastically changed the revenue matrix and IT investments models for businesses. Increasingly, businesses today are investing in IT systems to unlock the hidden value of data — not only the internal business data, but the vast quantities of social and IoT data to gain deeper insights into consumer patterns and thereby gain a competitive edge. Geospatial technology has become more accessible due to its online usage for mapping, directions and decision-making. Cloud computing helps many geospatial users to avoid the expense of hardware and focus more on their ability to manage and process data. The ability to manipulate and process data in a shared high-performance computing environment allows a user to access data from any location. Another tremendous change is the rapid rise of mobile offerings. Now, users can access smart apps that tap into data and analytics to support specific workflows. Today, more than ever, businesses are expecting their employees to work on the go. Whether they are at a client site, travelling, or working from home, employees are increasingly relying on mobile devices and the pervasive computing enabled by the Cloud. Businesses, therefore, can no longer rely on individual products and are increasingly focusing on seamless system integration and workflow solutions. Integration through Web Services, such as OGC Web Services, make the data accessible to address spatial information needs over global networks, including mobile devices. Geospatial technologies entering the IT mainstream There is overwhelming evidence that integration of geospatial systems into mainstream business and e-governance processes leads to enhanced effectiveness. With the proliferation of Web-based geospatial data and GPS-equipped smartphones, we are clearly seeing these restrictions and limitations disappearing and geospatial technologies entering the IT mainstream. Rolta has been at the forefront of developing intellectual property that integrates spatial and non-spatial data from disparate sources. Our endeavor has been to build industry-specific solutions that combine vertical domain knowledge with geospatial and IT expertise. These ‘productized’ solutions can be rapidly deployed for a faster RoI. Fast turnaround time is the deciding factor impacting business bottom lines. Companies cannot afford to reinvent the wheel. It is, therefore, natural to have synergetic collaboration and partnership to stay ahead of the competition. As I mentioned earlier, 80% of the data has a locational context and clearly mainstream IT companies, such as Amazon, Microsoft, Apple, Google, etc., are enriching their products and services with locational capabilities. This trend will undoubtedly increase the proliferation and adoption of geospatial technologies. Rolta has moved up the value chain from geospatial data creation to the development of geospatial fusion and geospatial analytics. Geospatial technologies will become an integral part of modern business solutions and decision-making. K.K. Singh Founder-Chairman & CEO, Rolta Geospatial World • January 2016 • 23 Industry Outlook Satellite Industry is Moving Toward Complete Workflow Solutions T he satellite industry is not functioning in isolation anymore. The industry is fast realizing that we are part of a solution for a problem we are trying to solve for a customer. Not only DigitalGlobe, but the whole satellite industry is moving to provide complete workflow solutions to customers. DigitalGlobe has taken the approach of Golden Circle — which is outside-in rather than inside-out. Tech companies typically are inside-out. Their traditional approach was: ‘I have this picture, what can we do with that?’ Instead, we can actually turn that around and ask the question, ‘Who is your customer?’ Once you know who your customer is, the next circle is what is the problem that they are facing. Once you know the pain point, the next circle is how do you address that problem. Finally, the outer circle happens to be what are the technologies that could solve that problem. It could be imagery, Dr Kumar Navulur Director, Next Generation Products, DigitalGlobe 24 • Geospatial World • January 2016 platform, algorithms… they need to be all brought together to solve the problem. We don’t need to reinvent the wheel, but collaborate DigitalGlobe has a strategic initiative called Geospatial Big Data & Analytics. The reason we call it strategic initiative is because we realize we need partners to make the solutions. When we talk about geospatial Big Data, our strength happens to be only part of the content — the imagery. And we can also derive some information layers automatically from the imagery. There are three types of services — platform as a service, content as a service [having data easily accessible and easily manipulated by the end users] and software as a service or tech as a service. This is where I see lot more collaboration with software image analysis companies. There is no reason for DigitalGlobe to reinvent the wheel. Our customers are not investing in IT. Instead, our job is to solve their problems with technology. In this case, we will put our imagery next to the advanced algorithms of a software company and bring together a solution for the customer. At DigitalGlobe, we believe in partnering — whether with people in midstream or upstream industries — and educate them on what is possible with imagery and our geospatial Big Data to deliver new solutions to those new industries. Whether it is the GIS industry or the imagery industry, we need to make all of it bigger. The way to do that is for us to be on the stage as service providers to both private and government sector. For this, we need to partner with deep subject matter experts. This is an industry in great need of expanding its horizons. The good news is that we are already doing it. Industry Outlook Customers Want Products that Make their Lives Simpler A t a recent IBM sponsored technology event in New York City, a senior executive started his briefing with a statement from an IDC Futurescape report: “One third of the Top 20 companies will be disrupted over the next three years because they are not moving fast enough.” Rapid change is the new normal in a global and flattened business landscape where disruption can occur from three people working out of a garage or when a large company decides to move into an adjacent market. For instance, Google’s recent decision to compete with Amazon Web Services for ‘on-demand’ supercomputing resources. Great products that meet mission-critical needs Customers want products that make their lives simpler. Services, system integration and workflows are all necessary components of a solution, but great product design can move those elements to the background. Think about the navigation system on your phone or in your car as the product you use every day to get from point A to point B. As a user, you just want to Stuart Blundell Director of Strategy & Business Development, Harris Geospatial Solutions 26 • Geospatial World • January 2016 type in the location and let the magic happen. The complexities of GPS triangulation, routing algorithms, and wireless services are all involved, but they stay hidden in the background. Using that paradigm, our focus is on building great products that meet mission-critical needs, whether it be secure network communications or weather satellites, and that satisfy that innate need for customers to make their lives simpler. Great product designs will always trump an alternative solution requiring extensive system integration. Hardware is and always will be key in delivering a “platform” for developing higher-margin software and related information services. Similarly, advanced space payloads and related ground processing infrastructure serve as a platform for geospatial Big Data analytics. The economics of moving data, processing and hosted services to commercial Cloud providers has become an accepted norm. The pace of change in technology in this space is phenomenal. Key drivers behind the concept of ‘sharing’ are open source Cloud platforms as a service [PaaS] providers, such as Cloud Foundry, as well as the real shortage of skilled workers in the areas of cloud computing and data science. There are only so many of these people in the job market and the competition to hire them is intense. My hunch is the concept of “sharing information infrastructure” will continue to evolve around these principles of open source PaaS and commercial Cloud services from major heavyweights. Increased Accessibility and Democratized EO Data A s a young company, we are well accustomed to pivoting and reinventing ourselves, sometimes by design and sometimes of necessity. So, we are yet to experience a time when innovation cycles have been long. That said, about two years ago, we assembled a talented R&D department to help keep UrtheCast ahead of the curve. It is quite possible that R&D has never been more important than it is at this time in history, which is probably true for the entire tech sector. There is an expectation that organizations will continually outdo themselves — and fast. This has been part of UrtheCast’s culture from the beginning: from our cameras on the ISS, to our Deimos satellites, to the development of our constellation of 16 optical and SAR satellites, to the development of our Web platform. Customers still want data Increased accessibility, fewer barriers, and democratized EO data — these are key elements in UrtheCast’s mission. From the start, it has been our aim to make this data more accessible to developers, so they can go on to build tools to address big monitoring needs: deforestation, refugee crises, environmental change, impacts of natural disasters, urban growth, natural resources, and so forth. We recently announced that the UrtheCast Developer Platform is now public. Pulling satellite imagery from a variety of sensors at different resolutions, developers can leverage not only our growing suite of camera sensors, but also third-party data. Imagery, video, and metadata are extracted from these sources to provide developers with an increasingly rich archive. Staying true to our vision of democratization, free access to the platform is available. These APIs are built on open, modern Web standards, and there is no need for people to be geospatial experts. It is a bit of a cliché in earth observation to say that customers want complete information solutions and services and not data or products, but, I think if we are honest with ourselves we as an industry have some ways to go in really achieving this in a comprehensive manner. Like others in the industry, this is our also part of the vision at UrtheCast, and something we are working toward, but frankly we are not fully there yet. Ultimately, our strategy to do this is through the API-based platform we are building, allowing customers to access our data directly or through third party developers’ apps that use our platform. Sharing information infrastructure It was inevitable that data sharing would become the standard — Cloud technology, the ubiquity of the Internet, and the need to solve increasingly complex problems paved the way for the amount of information we see shared today. In many ways, the space industry is coming late to this game. I think it is going to be fascinating to see how quickly and effectively the industry adopts and adapts such trends in the coming years. Wade Larson CEO & President, UrtheCast Geospatial World • January 2016 • 27 Industry Outlook Customers are Moving Hybrid Model Toward a A lot of people are familiar with large, expensive and great capabilities that exist with traditional satellites for geospatial data collection. They are, however, not very familiar that they can do great things with small satellites. As we keep up with technology, today we can do good things for a fraction of the size and cost. Many of these innovative companies have used advances in consumer electronics, automotive grade electronics, changes in manufacturing processes, materials, changes that occurred and continue to occur in computer processing, communications architecture, increasing bandwidth, etc. From data to complete analysis The US government today is looking at two ways to get their data. The NGA recently released its commercial GEOINT strategy, in which they talked about two ways of buying geospatial intelligence information. One is the traditional way — where the systems provide bits [0s and 1s] directly to them and they will work with that information. Another way they intend to do is to have a service where a company like Leidos will do the analytics using disparate sources of imagery, other sources of data, look for changes, look for interesting things that the agency would be interested in and alert them about the changes. Partnerships, collaborations Leidos is fairly well known in US government and industry as a company that leans forward and looks forward to new partnerships. We are always looking for small companies that we can bring onto our teams. We also mentor many small companies. First of all, we don’t have all the solutions. We are about integrating many good solutions into a greater solution. This is achieved by collaborating, partnering, bringing together new ideas, different people and companies. The other aspect of this is to help the users collaborate. We work with all the military, intelligence agencies, homeland security and other customers. We have an understanding what the customers are doing and sometimes what we end up doing is helping customer A to know that customer B has similar issue/project and if the two of them work together, they could probably do something greater together. Robert Zitz 28 • Geospatial World • January 2016 Senior Vice President & Chief Architect, Leidos W hat we are driving right now is the connectivity between the digital world and the real world. Whenever you need to create a new infrastructure, any plan is only as good as its assumptions. So, you need to know what the reality is. We are driving reality capture solutions by creating new ways to provide these realities to the digital world. Over the last few years, we have driven mobile reality capture quite dramatically. We fly in and build 3D cities for the ultimate app. While mapping street assets, we put our equipment on a car. We have ground penetrating radar technology, which can make the invisible visible, like the pipes in the ground. All these technologies help us to mirror the reality into a CAD file. We are building a bridge for our customers to move from traditional surveying to using extensive software and sophisticated instruments. We are binding our existing technologies of scanning into new software to make the workflow easy. Our reality-capturing technologies can decipher everything from traffic lights to the lines on the road, which then you need to maintain in preparation of the autonomous cars in the coming decade. Workflow needs to be automated in real time The cars need to know whether these assets are in place or not, and it is our job to make sure they are at the right place on the map at the right time. But, the workflow also needs to be automated and integrated in real time. When we Bringing the Real World into the Digital World capture data in different ways and amass it, we give the industry an opportunity to create new products. Transformation is not a new force at Hexagon. We have been transforming this market even before 1921. In the beginning, it was optomechanical and then it was optoelectronic, followed by the industrial revolutions. Lately, we are witnessing integration of software workflows into connected devices. Today, the process starts with the software on the sensor. You can get CAD files onto a total station and work on them. The other side to this transformation is to interlink these with other software. Since we live in a digital world, you cannot neglect the path from raw data to the application. If you do not link this information to the hardware, even the best software would not be able to keep you from losing information. It is only when you create a seamless link between the entire process that you can bring the real world into the digital world. Juergen Dold President, Hexagon Geosystems Geospatial World • January 2016 • 29 Industry Outlook Product-Technology Lifecycles Businesses O Shrinking and Reshaping ne of the most reflective changes in the recent past is the dramatic shrinkage of product/technology lifecycles. The collapse of lifecycles means that replacing a product or service line every two years is becoming the norm across many industries. Furthermore, if a business is not quick to introduce a product to the market, it risks launching goods that have already been superseded by competitors. This changing environment means that accurate market demand planning and forecasting has never been more imperative. The key to this is the innovation that enables organizations to quickly and effectively manage operations and gain a greater perspective over the entire supply chain. A similar example is the Internet of Things (IoT). The Internet has revolutionized the world in a big way. IoT will create significant disruption as well as significant opportunities in the coming years as companies create products and services that fit into a new world of connected devices and ecosystems. Solution-centric approach Solution has become the end of the value chain and a company cannot give a solution with just hardware alone. Trimble and Hexagon are the best examples here, which are primarily hardware companies, have major revenue sharing from software business now. As a part of innovation, versions of software being replaced by apps. Though they are not charging directly, they are teasing users with their products and solutions on their apps. Through freely available apps, the companies are aiming to increase their footprint in terms of outreach. Though the 30 • Geospatial World • January 2016 companies are shifting to software and apps, cost still remains a big challenge which is a hindrance to the growth of the industry. Mainstreaming of geospatial Mainstream IT players, like Amazon and Google, are venturing into the location sector for competitive advantage. The intention is not to develop geospatial products or solutions, but a need for new technologies to enable a seamless delivery in their own field. Ground-breaking acquisitions and mergers have been taking place in the industry as a way for businesses to expand horizons and gain competitive advantage. This is true for Apple’s acquisition of a US startup, Mapsesnse. Traditionally, GIS services were handled by specialist players in GIS while the IT services companies handled all other relevant IT systems. With the advancement of GIS across industry and user friendliness involved in viewing any relevant data on top of a map with right coordinates, the business users in the industry are starting to prefer any GUI or MIS report to a GIS map. This trend is making big companies develop in-home GIS groups to enable GIS across the business units. B.V.R. Mohan Reddy Excecutive Chairman, Cyient; Chairman, NASSCOM Location is an Important Factor in All Kinds of Data I ndustries are evolving to acquire end-to-end capabilities as part of a natural evolution. Desktop geospatial industry is evolving to enterprise and Web and is also providing capabilities to APIs. The applied use cases toward next generation businesses are also evolving, and some of them have been very disruptive to traditional industries. This trend, known as the consumerization of GIS, increasingly requires non-GIS experts to build and support location-based applications and services. Eventually, either specialized businesses will acquire horizontal technology stacks or geospatial companies will acquire further capabilities in vertical solutions. interoperability are the next big thing. Take for example, Big Data and Hadoop platform integrated with LI APIs. An agile and flexible platform can help organizations extract more value from their data to drive greater profitability, reduce risk and deliver more personalized and contextually relevant customer experiences. To really make sense of any data, location is an important factor. To do the analysis in location context, you need to geocode, route, geo-fence and run spatial analytics. The biggest social media companies could end up becoming the largest geospatial companies, as they are capturing billions of data points centered around their users. Open ended architecture and interoperability Real-time location data a disrupting technology The next wave of value creation would come from companies with solid technology platforms. APIs and microservices are becoming the core plumbing for the next set of architectures and workflows. And architectures that are fairly open-ended and are open to partnerships and Real-time location data became a disruption to the traditional taxi business model all over the world. Take Uber for example. Disruption to commercial taxi businesses via mobile apps is fundamentally dependent on location, where you are, where you want to go, what is real time weather, traffic, what is demographic location data analysis. To that business, location data becomes crucial to integrate. Similarly, e-commerce companies who are the next big trend are going mobile-first because Manish Choudhary VP, Global Engineering, & Managing Director, Pitney Bowes Software India they realize how to overlay their customers on a map at a particular place and time and draw inferences for shipping and tracking. These two non-traditional players have either acquired or partnered with geospatial companies to solve their critical problem and the same would apply to any data company. Geospatial World • January 2016 • 31 Industry Outlook A Successful Model Extends Verticallyand Horizontally both T here is no doubt that the product development cycle continues to be accelerated. Companies can no longer provide by themselves the compelling end-to-end solutions that the market is looking for. The challenge therefore lies is in understanding where one clearly adds value and has a sustainable competitive advantage and focus one’s efforts there, and then find the right partners in each target market to offer clients a superior complete solution. Given the high risk associated, this trend is particularly challenging for companies like Teledyne Optech that have always been technology innovators and market creators. Our organization has adjusted in several ways: we have moved away from pure ‘technology push’ innovations and have been working closely with forward-thinking clients and development partners to reduce market and technology risk and make sure that we get it right the first time. Software is the main driver Software has undoubtedly grown in prominence, as is expected in a maturing market. Today, clients are primarily interested in the data products generated through software at the end of the value chain. Software is also the main driver when it comes to productivity, a key decision factor in maturing markets. As a result, we have rebalanced our development team to reflect the growing importance of software and have adopted a systems approach that improves hardware and software integration. We are also recognizing that the value chain extends beyond Teledyne Optech’s core capabilities and knowledge base, and so we have been partnering and integrating with software companies that have established themselves as leaders in specific market verticals. Nonetheless, there continues to be a great deal of innovation at the hardware level, which then open the door for new markets and software-driven end products, or enable new levels of productivity never before achievable. Cost-effective and intuitive workflows We are working on both hardware and software fronts to make data acquisition, data processing and end-product generation as cost-effective and intuitive as possible. With features like PulseTRAK and SwathTRAK, our latest airborne LiDAR sensor, the Galaxy, has dramatically increased sensor efficiency and reduced operating costs, while totally eliminating the complexity in project planning and data collection. In the same way, our LMS software is now enabling the user to process LiDAR and imagery data through the same workflow and is optimizing the processing algorithms in order to achieve survey-grade accuracy even in very difficult environments. Michel Stanier COO, Teledyne Optech Industry Outlook As Solutions Become Complex, Collaboration Becomes Imperative R apid changes in product lifecycle are re-shaping the business by heavy competition and low cut off rates. In this evolving scenario, it is necessary to keep pace with market trends, and keep innovating and searching for faster and more efficient solutions. Organizations need to understand the sustainable products in their bag. They need to critically analyze the drastic change of demand to maintain their product lifecycle as there is rapid advancement of technology. This shift in demand is natural and expected. As solutions become complex and specialized, the need for implementers and integrators became imperative. Since clients expect endto-end solutions under one roof, the business ecosystem is developing in a more collaborative environment to fight back the competition. One-size-fits-all philosophy doesn’t work Today we cannot rely on the one-sizefits-all philosophy and custom-make each solution we offer for each individual. As the clientele grows more mature and technologically enlightened, we are expected to churn out high-quality solutions with niche technology at optimized cost. For instance, we at ADCC collaborated with domain-specific solution provider for BPAS and RAMS 34 • Geospatial World • January 2016 to compete in the market and enhance our opportunities. We also actively collaborate with industry giants like Esri, Intergraph, Bentley, DigitalGlobe, etc. For overseas business, we collaborate with local companies to increase our credentials, to reduce cultural and language gap, and to understand the market. I believe there cannot be a standard model for collaboration, we need to follow due diligence and be customer-centric in our approach to build solutions and must mutually fill in the gaps.The trend of sharing information infrastructure is a positive step toward better business environment. This has enabled specialized solution providers to come together and create precise offerings that are better suited to the client’s needs. Information sharing also plays a crucial role for organizations to execute quick and cost-effective solutions. Governments are coming up with various solutions, such as SDIs, to bring all data on a common platform for them be accessible by different organizations. This will reduce the cost of data preparation and also repetition of data creation. The best part about geospatial solutions is that it not just adds value to your offerings, but it also gives us smart tools. Amit Somani Joint Managing Director, ADCC Infocad Innovations are Transforming the Industry O ne of the trends we are witnessing today is of companies building their businesses around geospatial technology as well as data. Uber would be a good example of this. Companies are innovating to improve the quality of maps, to quickly map new areas, and develop a new technology for pin-pointing location more precisely, both inside and outside. So, there is a lot of innovation which is being fuelled by these new applications. This has also created a need for the industry to broaden its definition. We need to look at ourselves as more than just horizontal providers of technology. The industry has transformed into a bigger player solving consumer and/or business problems. Mainstreaming of geospatial Because of evolution in the startup space and competition, even large players like Google and Apple are trying to make their platform better. All this has created an opportunity for companies like RMSI. If the big players have to ramp up or improve their data, they need to reduce the time to market. As a whole, this is creating opportunities for the industry. The apparatus fuelling innovation is the sharing information infrastructure. This has opened up dialogue about Big Data, the benefits of which are yet to be realized by the geospatial domain. If all the data from various countries and different regions were to combine with all the other data from sensor systems, imagine the kind of insights we would be able to get. We can create new applications to solve complex problems using that insight. Information sharing is also important for small and medium enterprises because it enables them to compete on a global scale without being weighed down by upfront investments. So, in a way, it has unlocked the startups, as well as the mid-sized businesses, and given them the opportunity to go global. Traditionally, geospatial has been a very specialized field with specialized skills. But, all these activities have mainstreamed geospatial to such an extent that companies like Anup Jindal CEO & Joint Managing Director, RMSI Google are openly saying how their strategy revolves around maps. Mobile platform is important This year, a very fundamental shift happened: the Internet access on smartphones took over the Internet access on desktops across the world. This means that, increasingly, more and more Internet access would happen on smartphones now, which would blur the difference between an app and a Web application. All your technologies need to work seamlessly on a mobile device; you cannot ignore the mobile platform anymore. Companies need to internally set up the capacity, capability and innovation of increasing their mobile outreach. Extensive use of mobile devices also means that companies can bring in new sensors. Soon, they will even start monitoring pollution levels in the cities. Tech Trends | Small Satellites Small sats have come into their own in Big Boys’ realm S pace technology used to be a big boys realm. Big budgets, big facilities, big teams and big time schedules summed up these efforts. Usually, these were government laboratories, as in the case of India, or government-funded projects in big companies, as in the United States of America. The users also were big government departments and big corporations. However, this big approach had limitations. Beginning with ERTS (now Landsat), the initial euphoria of seeing large swathes of land in one scene was soon tempered by the amount of information that could be drawn from these scenes. Large swathes of farmland were visible, but individual crops could barely be identified unless they covered large areas. This required higher resolutions, but swath got limited and so did the revisit periods, making the monitoring of the crop growth difficult. Other applications like defense, urban and disaster manage- ment demanded high resolutions, shorter revisit times and faster data processing. Multiple satellites, pointable sensors, agile spacecrafts solved some of these problems, but always at the cost of some parameter. Multiple satellites increased the cost, pointable sensors and agile satellites resulted in limited coverage. This is when small satellites came into their own. They cost less, both in terms of the satellites and launch costs, and, therefore, a multiple set of such satellites could be arranged to fly in formation as a constellation to provide both wider coverage and higher revisit. However, the time from imaging to data in the hand of the user remains a problem because of the sophisticated preprocessing. Advances in electronics are now enabling the development of nano and microsatellites. Also, these can now be designed and made in simple laboratories, as can be seen from the large number of satellites designed and built by students. 36 • Geospatial World • January 2016 Data and data processing is now more accessible through the Cloud. It is not only students, but, with the maturing of technology, many industries are developing microsatellites with seven-year life and ‘appropriate’ risk-reward that satisfies venture capital providers. Nanosatellites may lack in sophistication, but they meet the need of frequent revisit and fast turnaround of data which suits many critical applications, like defense and disaster support. Josh Alban, Vice President – Sales, Planet Labs Forcing the industry to move faster Instead of just looking for highly precise and high-resolution pixels, we are opening up a new league of data in and around time resolution — data based on temporal resolution. Nanosatellites are forcing the industry to move at a fundamentally faster pace. If you are not taking 5 to 10 years just to build a satellite, you can innovate on the core technology more quickly. That enables a company to serve better and more interesting datasets to not only existing customers, but also to people who know very little about geospatial information. Today, we are benefiting from the years of investment the consumer electronics community has put in nanotechnologies. In many ways, our small satellites are producing superior datasets compared to very large and expensive satellites. Opens up the space for small countries A lot of small countries with limited budget are investing in small satellite technology. It is a lot cheaper application to invest in when compared with starting a full space science program or developing geo capabilities and satellite communication. Earth observation is the wisest start in this industrial process with the potential of developing more extensive programs later on. We are seeing countries such as South Korea or even Nigeria and Algeria who started off with this technology and now have a diverse EO program. We expect other countries to follow this course. Low capex model and competitive pricing We have seen an unprecedented level of investment — $415 million — coming into the business in the last couple of years. The low-capex model is attractive to investors. There is, of course, a risk factor. But it is On the back side of all of this data creation, there is an analytics community which is frustrated with relatively shallow global spatial datasets, and is ready for our new temporal datasets. In the last five years, this community has made rapid strides in the fields of machine learning and Big Data concepts. We look forward to see what they will do with our data. Frequent launches a plus In a more traditional cadence of launching, failures would be a very big problem [Planet lost 34 satellites in launch failures in 2015]. But, we have launched nine times with a failure rate that is quite low. The fact that we launch more frequently means that one failure would not impact the broader evolution of the constellation. important for the industry that some of these projects are successful so that future investments are not affected. Microsat data is priced very competitively. They also have the capability of launching a vast number of satellites in one go. Small sats answer the questions of temporary revisit, bringing it down to hours, rather than days. We do not necessarily expect it to tap into the high-precision defense markets. The data has the potential to open up completely different market areas, such as, in market and business intelligence, support to financial communities, insurance, etc. Some of the monitoring-style applications will see increased competition and the trend could prove to be disruptive. Adam Keith, Managing Director, Euroconsult Canada Geospatial World • January 2016 • 37 Tech Trends | Small Satellites Big players are here to stay, but small sats will expand the market In many recent cases, these embody a disruptive movement to provide rapid, low-cost data service to a much wider audience/market. The compromise is the achievable swath and resolution. Often times, nano and microsatellites might be quoted to have high resolution, when really it is the ground sampling distance that is being specified and not the actual resolution. You can’t break the laws of physics, and usually, a larger instrument is required for high spatial and spectral resolution and/or large swath. For these missions, the satellites will be larger. One can think of nano and microsatellites as expanding the market to include users of lower resolution data that would not otherwise have access to images from more expensive satellites. Nano and microsatellite approaches are also being used to reap the benefits of mass production for large constellations. These constellations increase the total volume of data available and reduce the latency of data. This again promotes market growth to include those who need low latency, quick revisit, and a more images in a short period of time. Quicker revisits, lower cost We need to produce satellites that are not top performers in favor of quicker revisit, higher data volumes and lower cost. This makes imagery accessible to a larger, untapped market. The cost of satellites, ground stations, and operations needs to be kept low enough to allow business models to close. Otherwise, commercial exploitation of satellite imaging will not be successful. The traditional big players, like DigitalGlobe, will likely still have a role with the high resolution, high performance missions. However, these large satellites will not be amenable to mass production and commercial data services owing to their cost. Dr Robert E Zee, Director, Space Flight Laboratory, Microsatellite Science and Technology Center, University of Toronto Institute for Aerospace Studies Real disruptive business segments will emerge Nanosatellites are great for a few applications and they are simply not right for others, like the more classical intel- or defence-related applications that need high resolution and high frequency. The business of EO imagery is very much government driven. But now, we are seeing a new B2C business developing on the top of this and that is fuelled by nanosatellites. This is the kind of demand that will develop with the platform itself. So, the classical segments will remain there and will grow steadily in the next few years, but the real disruption is going to come from the other side. And this would be possible only if there are innovations in Web technology for allowing the people to access this data and the data stream that populate this platform with counters. Affordable and democratic data One of the key points of being able to expand the market is now the availability of reliable data. But this data has to be available at the right price and with a proper frequency. So it is a combination of three factors — the data quality, the price and the refresh rates. Further, the way this data is distributed is becoming more and more innovative, so we are able to access EO imagery not only faster and cheaper, but also in a more democratic way. It not only allows more users to access the data, but also access it differently to be able to process it in real time. Fabrizio Pirondini, CEO, Deimos Imaging 38 • Geospatial World • January 2016 Makes a great business sense The Silicon Valley venture capitalists The trend of microsatellites is catching up because of several great business reasons. Essentially, these can boil down to technology and availability of capital. It is now possible to manufacture satellites that are just as reliable [depending on the application] for a lot less money. In our case, we have been through three separate designs for a satellite, and we finally arrived at the one which has the appropriate risk reward in terms of what it would cost to fly it. And we have gone down dramatically in terms of the weight and the cost; not only what it costs to build a satellite, but also in terms of launching it. For each kilogram that you take off a satellite, you save tens of thousands of dollars in launching it. And we have built satellites weighing as less as 15 kgs. There’s a tremendous amount of technology innovation that is occurring in the sphere of engineering and manufacturing of these satellites, and PlanetiQ has benefited greatly from it. If I talk about the other driving force behind this trend, which is the financial one, you need to know what venture capitalists in the Silicon Valley would be able to finance. We are witnessing an explosion in the 3U satellite form because they are very inexpensive and are considered by some as ‘disposable’. We, however, have not taken that approach. We don’t view them as disposable; in fact, we have a seven-year design life on ours. I would say that we have gone up in the value chain by building a more robust satellite. But, the venture capital community is very detail-oriented and they want to weigh their risk-reward on the amount of money they are willing to put forward, and the amount it takes to build these instruments. Anne Hale Miglarese, President, PlanetiQ Why do we need to re-invent There is a lot of commercial opportunity in the 1-metre resolution space. We have to look at better and higher resolutions in this area. At BlackSky Global, we are deploying a constellation of 60 satellites to change how we look at the planet. So, instead of viewing 1 or 2 revisits per day we are planning 40-70 revisits per day. Ultimately, it is going to be very disruptive on how we look at the planet in a positive way because suddenly we can do things with satellites that were not possible before. Big satellite players like DigitalGlobe are the biggest leaders in the industry. They have 80% of the revenue, they have a great spark to capitalize the revenue and they really have the customer base. And they will continue to be the leaders 5 to 10 years down the line. Those who innovate will stay relevant Our business, however, is going to come from everywhere. Ultimately, it comes to three things — revisit rates, the price point for the products, and the user experience. As an industry, we can do much better in all these areas. As we bring the imagery prices down with time, we will be opening up new markets. Also, we need to focus on distribution partners and channels. We have to think about what that means to the satellite industry, the ground segment, the distributors, the analytics community... So, there is a lot of business opportunity throughout the value chain. Organizations that innovate will stay relevant in this market and those who do not, could be left behind. Jason Andrew, CEO, BlackSky Geospatial World • January 2016 • 39 Tech Trends | UAVs drones Take wings, but still bound by regulations U AVs and their cousins — the drones — have been in the news for all the wrong reasons. TV serials like Homeland have made them out to be evil destructive inventions. The fact is, they range from toy helicopters to those evil, destructive things we see on the TV. Somewhere in the middle are the peaceful ones, ranging from a quadcopter carrying a video camera and recording a wedding procession to more sophisticated ones which are stabilized and mounted with sensors like digital cameras and laser scanners for remote sensing. The use of UAVs for remote sensing fills a big need. They can be flown at short notice, and can cover small areas in very high resolution as the flying height is low. The cost of operation is also low and data is available almost immediately. For example, they have been used in the Nepal earthquake recovery management by students with a minimum amount of ‘flight training’. While digital cameras have been the sensor of choice till now, laser scanners are fast becoming popular. Does it mean that remote sensing from high resolution satellites will be replaced? The answer is no. In fact, neither will aerial photography become obsolete. UAVs provide a quick and easy solution and that is also their weakness. While they are good for single sorties over small areas, they cannot replace aircraft which can cover multiple large areas in a single sortie. Further, modern aircraft sensors are complex and yield multispectral data which can be used for multiple purposes. Satellite remote sensing can cover even more areas, which gives a degree of synopticity not available from UAVs or aircrafts. UAVs are limited by regulations of where they can be flown. These regulations vary from country to country. In some countries, there is a complete ban, while in other countries, 40 • Geospatial World • January 2016 the operation is limited in height and to areas away from aircraft routes and range is limited to line of sight between the operator and the UAV. Despite these limitations, the use of UAVs is expected to grow particularly as new lightweight sensors are being developed exclusively to be used on UAVs. It is about democratization of aerial imaging and data itself Vast majority of UAS will continue to be imagery based UAS provides accurate aerial imagery — previously controlled by government, aerial mapping and satellite firms — accessible to all. In that sense, it is less about a technology or interesting drone concept and more about the democratization of aerial imaging and data itself. Aerial imaging is one of the most powerful forms of geospatial information available, across a vast range of industries and applications. UAS is playing an increasingly important role, and now complements traditional aerial survey, mapping and satellite data. Sensor development, and the decrease in the actual size of sensors while keeping or improving their functionality, is the main reason the UAS market is able to exist today in its current form. Laser scanning is less important than imagery in the UAS market, given the advances in photogrammetry we have seen, especially in the recent years. Scanning will continue to be a niche sensor, but it is unlikely to be widespread in the UAS market given the cost and complexity. The vast majority of UAS are, and will continue to be, imagery-based. Beyond visual line of sight is interesting as it would allow corridor mapping by UAS. There are still significant technical, safety and regulatory concerns that need to be addressed before a UAS with few or no redundant systems could fly in such conditions. Until then, it’s an interesting industry to watch. Todd Steiner, Marketing Director – Imaging Business Area, Trimble New way to safely and efficiently collect spatial data UAVs are continuing to push the paradigm shift of moving from ‘point-to-point’ data collection to a ‘continuous representation of reality’ — capturing all information in a scene at once and distilling that down to required spatial data (points, lines, polygons, attributes, orthomosaics and digital elevation models) on the desktop. In addition, we shift from ground data collection to aerial, which provides a new way to safely and efficiently collect spatial data. Data can be collected in areas that may not be accessible in any other way in a fast and efficient manner. Users can access the data they need in a safe, easy, efficient and profitable way. While sensor development has lagged behind the pace of the UAV hardware, it is rapidly improving with the emergence of new options for both hyperspectral and thermal imaging. Current laser scanning technology needs to change dramatically, in both size, weight, and energy consumption, to be used on UAV hardware platforms. Good fit for core geospatial businesses UAV technology is a good fit for our core business, representing another valuable tool for our customers to increase their efficiencies and stay competitive. Agriculture is a significant market for Topcon with high UAV acceptance. And with our recent acquisition of Digi-Star, aerial mapping is proving to be a helpful tool for crop management, irrigation planning and determining plant health. Our aerial mapping solutions have also brought us into new markets, such as inspection and monitoring. I am sure we will continue to be surprised by the applications that arise for UAVs. Charles Rihner, Vice President – GeoPositioning Solutions, Topcon Positioning Group Geospatial World • January 2016 • 41 Tech Trends | UAVs The real challenge is to move into apps There has been an explosion in the UAV market, and analysts are predicting further growth. There is a lot of competition in the segment. However, it is not enough to deliver only the nice copter technology. I think we have the nicest copter; we have a customer who said it looks like a iPhone. But this is not enough. The real challenge is to move into the applications. Customers not only want drones, they also want core applications so that they can work with all the data which we can generate with the drone technology. And for this, we are developing a lot of applications for stepping in the surveying business. We now have the new functionality for high precision GNSS so that the customer can work very efficiently. They can speed up in their surveying process and this is then very helpful. value, and so we need applications with which you can move in the whole business. And this, I think, is the main success factor. It is also very important that UAVs are capable of adopting very flexible and varied sensors because every customer has a different application, a different sense of philosophy. This is our concept at the moment: you can adapt very easy, different sensor technologies, optics so we can cover a lot of applications. Create business space and add value You have to create business spaces, you have to convince the customers that you can generate and add Dr Herbert Machill, CEO, Aibotix Uncertainty over standards and rules The advent of UAV/RPAS as a low-cost platform for mapping and surveillance purposes has generated a revolution in platform development with an exponential number of configurations and sizes. While this explosion has created many options at very different price/performance points, it has also led to extreme fragmentation and uncertainty about standards and rules, ultimately resulting in much confusion in the marketplace. Given the payload limitations of most UAVs, imaging sensors have primarily been deployed based on their inherently smaller size and power requirements, and they have been challenging the scanning market by asking the question, ‘What is good enough?’ Platform development has been far and away more prolific than LiDAR sensor development. There are LiDAR sensors available that currently serve the UAV market space, but typically these sensors either only support general mapping requirements in terms of accuracy and measurement precision, or they are priced too high and remain outside the payload capabilities of most UAV platforms. At this stage, technology limitations and the absence of platform standardization prevent the development of a sufficiently performing yet low-priced LiDAR sensor that could seriously disrupt the engineering-grade survey market. Cost savings are undeniable UAV deployment as an alternative platform for low-cost data collection is attractive to many. Cost savings are undeniable in the context of single localized surveys. However, their adoption for multiple same-day collects of independent survey areas is impractical and inefficient. In these cases, low-cost manned platforms are still more efficient, and are not affected by the current legislative restrictions that govern the UAVs. Therefore, manufacturers must consider sensor development in the context of available platforms, required accuracies and the application verticals in which they choose to participate. Mike Sitar, Business Unit Manager, Airborne Mapping Solutions, Teledyne Optech 42 • Geospatial World • January 2016 Integration into national airspace Advances in smaller and lighter sensors Unmanned Aerial Systems (UAS) are changing the landscape of many industries and finding roles in almost every walk of life. As far as the spatial industry is concerned, the real advantages are to be found in the sensors. The lower cost of operation compared to traditional means of data collection has also meant that more data can be collected, integrated and cross referenced to create ever more detail of the subject area. UAS come in a variety of sizes with increasing payload capabilities. However, it is important to utilize the right technology for the given task. Developments in the integration of larger UAS into national airspace will give the spatial industry new horizons when it comes to covering large areas quickly and efficiently. UAS technology developed in the past few years,but it is the sensors that deliver the real results. The advances in smaller and lighter sensors, in conjunction with the developments in the aviation regulatory environment, has allowed UAS to be operated in all sorts of environments, delivering all sorts of information that only a few years ago could not be dreamt of. These advances (in photographic resolution, LiDAR and laser scanning technologies, even spectral analysis) all needed a very stable platform and have found a natural home with UAS. One of the biggest new advances that I see happening in the next few years is an increasing development of large UAS capable of operation alongside manned aviation, and covering large distances whilst carrying a wide array of sensors. Couple this with the anticipated advances in computing and the benefits will be staggering. Systems will be able to operate at considerable distances from base (with no need for co-located operators) providing data at very short notices in quantities not yet envisaged. Gary Clayton, Chairman, Unmanned Aerial Vehicle Systems Smaller, more efficient and all-weather UAVs provide an alternative, more cost-effective method for collecting data. They are smaller and, therefore, can be used when needed; rather than scheduling time and resources to use manned aircraft or satellites. They can often fly closer to the targeted area to collect more accurate data than possible with alternate methods. They can also supplement other data that is collected, so more information is available for analysis and reporting. UAVs can also be used in weather conditions that other aircraft cannot operate, thus ensuring that projects stay on target. The markets that rely on up-to-date, accurate and detailed GIS data will adopt UAV technology more quickly by having additional sensor payloads available. Users who want to get their jobs done more accurately and more efficiently will discover new applications. Sensor manufacturers need to catch up The UAV market and technology are evolving at a rapid rate, making it difficult for established sensor manufacturers to keep up with this pace. The challenge is that most sensors today haven’t been designed for unique characteristics or variation in the sizes of UAVs. Bulky and full metal encased sensors on the underside of a UAV can drastically impact the flight performance, communication link, reliability and safety of a UAV. Unless you operate a large UAV, these sensors can’t be used. That said, sensor manufacturers are now realizing the potential of the UAV market for new business and are starting to adapt their technology to integrate with UAVs. We are starting to see smaller, lighter sensors with the same or better performance than “off-the-shelf” sensors. Laser scanning tech is catching up. David Proulx, VP Product & Marketing, Aeryon Labs Geospatial World • January 2016 • 43 Tech Trends | Big Data The future is in analytics & interconnected Big Data W ith the advent of satellite remote sensing, Global Navigation Satellite Systems, aerial surveys, sensor networks, radar, and LiDAR, the volume of geospatial data has grown exponentially with data production crossing storage capacity in 2007 itself. To these conventional sources, add the unconventional sources like location aware devices, crowd-sourced data, RFIDs and IoT data sources and you have a huge data stream. In its report on Big Data, McKinsey Global Institute estimates that location data level stood at 1 petabyte in 2009 and had a growth rate of 20% per year. This did not include data from RFID sensors. It also does not include ‘dark data’, that is, data collected by researchers and lying in private archives. The UN-GGIM estimates that of the 2.5 quintillion bytes of data generated everyday, a significant portion of the data is location-aware. However, it would be wrong to create bins for geospatial Big Data and other Big Data. The key to Big Data is the need to look at all data in ensembles specific to applications. Thus, according to Prof Shashi Shekhar of the University of Minnesota, Big Data also includes data-intensive computing, middleware, analytics and scientific and social applications. In fact, the current trend of interconnecting different data sources through the Web will ultimately give rise to the Internet of Things, where every source will have its own Uniform Resource Identifier and geospatial data will become omnipresent. There will be over 50 billion such resources by 2020. The UN-GGIM report on future trends in data over the next 5-10 years indicates that Big Data is coming to the fore. It is of the opinion that the future is not one Big Data store, but a number of datastores interconnected and accessed through models satisfying different 44 • Geospatial World • January 2016 end uses. Big Data will require open standards to discover, access and publish data. The provenance of data will also become important. Big Data is increasingly finding importance in transport, logistics and disaster management which present the challenges of data volume, velocity and variety. Big Data is junk if you can’t understand it Everything that happens, happens somewhere. It’s the ‘where’ that provides context. In the words of Richard Saul Wurman, “Big Data is junk if you can’t understand it, but a map is a pattern made understandable.” Practically, everything has an explicit or implicit geospatial location — explicit when a latitude and a longitude are provided, implicit when an address is specific, like 380 New York Street, or implied in a mobile Tweet such as, “I’m eating a Chicago style pizza in downtown Manhattan.” Even if the location service on the phone is disabled, a natural-language processing analyzer can determine that Chicago is not the correct location of the Tweet — downtown Manhattan is the location, and that can be converted into explicit latitudinal and longitudinal values. What’s getting better is our ability to model and forecast our cultural, physical, and biological futures. The difference is real-time data. You have sensors pulling in information about traffic, noise, air pollution, water quality. GIS has always been about data, but now GIS is getting filled with streams of real-time data. We are able to integrate that data from different sources and analyze it against historical patterns to make predictions. So instead of simply telling you what current traffic conditions are, we can predict what the next hour of traffic will be like and tell you the best way to get where you’re going. Beyond the 3 Vs Big data usage is more than the traditional volume, velocity and variety (3 Vs); it is the realization that sometimes the 3 Vs just don’t apply. For example, imagine if another Fukushima Daiichi event happens — the sparse data from sensors in a remote village is not coming in fast, is not big and is very structured. However, the window of opportunity to respond to such an event is so small that we need new ways to determine whether that village needs to be evacuated, and we need certainty in our decision making. This is where geospatial analytics is most relevant in the form of Bayesian kriging with regression executed very quickly by nontraditional means. Mansour Raad, Senior Software Architect, Big Data Advocate, Esri Multi-model analysis There have been a lot of new innovations in the world of Big Data and Big Data modeling. There is a lot of data coming in. New innovations like self-driving cars are bringing in completely new kinds of real-time data. There are some buzzwords — real-time, in memory, relational table and rows, NoSQL key-value, Hadoop sharding, graphs — that are going around for new technologies. All of this comes down to the multi-model store that is doing the analysis of the data — whether it is coming from existing databases, materials coming from blogs, financial reports and data from the agriculture world. Then there’s real-time information and electrical monitoring, etc., like the Oracle Multi-Model platform and Cloud Data platforms One Shared Multi-Model Store or Multiple Independent Stores. It depends on the customer’s choice and supports any data type, any scale, on-premises or in the Cloud. Whatever is right for your company or your application, the ‘multi model’ platform can handle it. Re-purposing the data Because you have a lot of data and it’s complicated, you need to re-purpose the data. The data is already in a store, but different human beings will use the data in different ways. Now, environmental monitoring people, agricultural people have different terminologies for the same data. You can use different terminology to remap it into the terminology used by that particular customer. But you are paid to collect, store and model the data only one time. That’s the beauty of providing the semantic layer on top of it. Convergence requires access to different data model stores either write custom software or use semantics. Steve Hagan, Vice President, Database Server Technologies, Oracle Geospatial World • January 2016 • 45 Tech Trends | Mobile Mapping Proliferation in sensors driving mobile mapping O ne of the most difficult tasks for a city administration is to keep track of the construction and modification of buildings within its jurisdiction. While there are documents galore of permissions and payment of taxes, these do not include the actual situation on the ground. Satellite imagery from time to time can track changes but cannot be used to determine adherence to laws. Mobile mapping, which uses a 360-degree viewing camera system mounted on a car, can collect more useful data as the car travels down city roads. This data can then be analyzed offline for illegal constructions and modifications. Such imagery is precisely located using GPS and can be used for precise mensuration of building façades. Mobile mapping also provides the ‘street view’ in Google maps and other such sites, providing an augmented reality view to users. Apart from photographic cameras, laser scanners are also used for the same purpose. Terrestrial laser scanners can be used in mobile mapping for other areas, like recording mine activities, mapping and virtual reconstruction of heritage sites and mapping of geological and geomorphological hazards like landslides and rockfalls. Mobile mapping is also used for determining rights of way for future roads, pipelines, transmission lines and railway lines. For many of these applications where vehicles cannot be used UAV mounted scanners are adopted. Ship-mounted scanners are also very useful for mapping the coastal environment. These applications earlier required a considerable post-processing and the access to the results were limited. Now apps are being developed that provide the results on portable devices that can be used in-situ. Also, the ensemble of sensors is growing and now includes 48 • Geospatial World • January 2016 in-situ sensors and other devices like autonomous car sensors. Post-processing is moving to the Cloud and apps soon will be able to analyze the data for specific end-uses. Dr Harald Teufelsbauer, Product Manager, Mobile Laser Scanning, Riegl Diversity of applications is a huge plus Mobile mapping helps in time-efficient and safe surveying of large areas. The dense point cloud that is created from mobile mapping creates detailed and accurate acquisition of topography, façades, pavement, traffic signs, construction sites, and other minute details. Mobile mapping provides a precise level of detail that can be used for a variety of tasks, such as, engineering, modelling, and monitoring, and also for training and planning purposes in surveillance activities. Recent developments in autonomous driving require highly precise 3D road maps. Various automotive industries have already taken to mobile mapping to produce a new dimension of road maps. The diversity of applications that mobile laser scanning can be used for is a huge plus. Different levels of information can be extracted out of a point cloud acquired in just one acquisition campaign. Compared to traditional single point surveying where targets are mostly captured following Across-the-spectrum benefits Smartphones and tablets, with their sundry sensors — including GPS and other location services — significantly enhance the delivery of mobile mapping and location-enabled apps and services to a large audience. The cost for ‘good enough’ accuracy in commercial grade devices is significantly less in every dimension — device, training, sensors and applications. In addition, the apps delivered to consumer devices today are generally more advanced in usability and integration than the software on proprietary devices. Location-based services and apps integrated into routing and logistics workflows have reduced the cost of getting things where they need to go. Whether it is iron to a steel mill, or flour to a bakery, or beans to a café, companies and enterprises are benefiting from mobile mapping. Ironically, it is some of the traditional survey companies and a few government agencies that are benefiting the least — the former not being able to navigate the innovators’ the requirements of one specific application, the immense level of detail in laser scanning point clouds allows multiple uses of one dataset for a wide field of applications. Growing demand for hybrid systems To a certain degree, the accuracy of mobile laser scanning systems is dependent on the availability and quality of GNSS, which can be a limitation for certain applications. In GNSS-denied areas, terrestrial laser scanning is the complementary solution to create highly precise point clouds. Of course, there are a few challenges that we see in the wider uptake of mobile laser scanning. With increased scanner performance and high level of point cloud detail, further development in smart data processing will gain importance in the field to make businesses more efficient. There is also a growing demand for hybrid systems, such as, terrestrial to mobile systems or mobile to airborne systems, without losing performance for special applications. dilemma, and the latter being hamstrung by moribund IT policies and glacial adoption rates. Too strictly tethered to Web services One of the limitations of mobile apps is being too strictly tethered to Web services. Part of this is driven by technology — the computing resources behind say Google or Amazon Web Services are more powerful than a smartphone. However, vendors who don’t want any separation between them and customers are driving much of this. This has a trickle-down effect, where applications don’t work as efficiently in low-bandwidth environments or even work at all when off the grid completely. George Demmy, Chief Technology Officer, TerraGo Geospatial World • January 2016 • 49 Tech Trends | Mobile Mapping A trend from a solution standpoint Mobile mapping is driven by the need to efficiently collect 3D position, condition and visualization data across networks, corridors and high-value sites. Traditional surveying is not suited to cover data collection of 100s or 1,000s of kms of roads, pipelines or powerlines efficiently. The move toward making mobile mapping technology and data accessible to more users is the main industry trend from a solution standpoint. This includes: intuitive touchscreen, ease of use, dramatically lower costs coupled with higher functionality and imagery as the ideal for both human and machine interpretation of mobile mapping data, and Cloud infrastructure and ecosystems to process, host and publish data. The main challenge with mobile mapping is the cost to collect data, like the road network of a state or a country — mobilization, salaries, vehicles, etc. Perhaps, as cars increasingly have on-board cameras and sensors for semi-autonomous and autonomous driving, they will become a source of data and the industry will re-focus on software, apps and delivering and publishing data and information. In a long term, mobile mapping, as we know it today, could cease to exist, and evolve to being something commonly used, whether actively or passively, by almost everyone. Apps to be a growth area Apps allow customer problems to be addressed and solved, like pavement marking detection or volume calculations on mines. This is a new trend in mobile mapping and will be a growth area in the future, allowing more people to participate in this market and provide a broader range of solutions, especially to more specialized applications. Moving forward, the importance of apps will be coupled with the ability to process, share and publish data in the Cloud. The amount of data collected and the resulting logistical challenges to transfer data from the field to the office, and then analyzing the data within a time-sensitive framework, remains a challenge. Eric McCuaig, Market Manager, Land Mobile and Aerial Mapping, Trimble A paradigm shift The emergence of mobile mapping technology has brought about a paradigm shift in moving from “point-to-point” data collection to a “continuous representation of reality” — capturing all information in a scene at once and distilling that down to required spatial data on the desktop. The biggest advantages of collecting asset data with a mobile mapping system are speed and efficiency. A mobile mapping system will collect all roadside assets while traveling at normal vehicle speeds. Specific points of interest can then be taken from that dataset on the desktop at any time, eliminating field re-visits. An additional benefit is safety: mobile mapping from a vehicle allows the user to safely collect data and eliminates the need for fieldworkers to work in hazardous conditions on highways and roads. Speed and ease of use Mass data collection for utilities, road mapping, construction, railway mapping, and mines represent common applications. Mobile apps are not only increasing the speed in which that data can be accessed and used, but also the expectations of how quickly the data should be collected and used. Mobile mapping systems offer a good RoI when an agency has a large, ground-accessible area to be mapped. For agencies that just need a small amount of field mapping done, purchasing a mobile mapping system may not be the right choice. Accessibility is a major limitation in terms of the technology. There are occasions when the mapping vehicle cannot get to every location where a user requires data. New technologies such as unmanned aerial systems can be used in combination with mobile mapping when data is needed in these areas. Dave Henderson, Director – Geospatial Solutions, Topcon Positioning Group 50 • Geospatial World • January 2016 High E ffic ienc y B et ter Per for ma nce http://www.kqgeo.com/en Email: ank e@kqgeo.com M alaysia O ffice: Level 41, V ista Tower The Inter mar k, 348 Jalan Tun R azak, 50400 KL KQ GEO Technologies Co., Ltd B eij ing GIS S of t ware Wuha n Industr ia l S o lutio ns G u ang zhou Ku a la Lu m p u r Cu s tom ization Su r vey Ins tru m ent s Tech Trends | Apps apps the new way to go, Thanks to mobile devices A pplication is, what we were taught, the use of data to create meaningful information using the tools of our trade. An example is the analysis of remote sensing data using an image analysis software and ground truth to create a meaningful map, like a land-use map. Since then, applications have moved beyond such simple tasks to more complicated ones, like how to align a new road. Along the way the abbreviation of applications, apps, has become a word by itself. So what are apps? The growth of apps has been driven by the evolution of mobile devices, like tablets and smartphones. An app usually invokes complicated functions, usually somewhere in the Cloud, to give the user a simple answer. What is new in the world of apps is the professional app that cater to users from the fields of science, engineering, defence, homeland security, administration and business. Apps have become very important in the field of C4ISR as well because soldiers have become sensors and receive on line instructions from their command base. Police, firefighters and rescue teams require apps to 52 • Geospatial World • January 2016 navigate areas under threat or damage. Since professionals do not have to go back to their desktops for analysis, this enables in-situ decision making and viewing the results in real time or near real time. Apps are platform and software neutral and have to follow open standards to be acceptable in the marketplace. This is an excellent opening for independent developers. Dirk Gorter, Director of Product Management, Esri Leverage the power of Web GIS Traditional GIS users are empowered with rich content and processing, but a whole new class of apps is making all this information more useful and timely to the users. These apps allow users to integrate, visualize, and analyze geospatial data of all types. All this comes together to extend the traditional role of GIS as not only the system of record, but also as the system of engagement, providing the basis for understanding, communicating, and collaborating across the organization and beyond. The apps in a Web GIS range from rich desktop apps to browser-based Web apps to native mobile apps. The apps world has moved to using focused apps for specific tasks. Mobile apps are a good fit for people collecting data out in the field. They need apps that work well, either connected or disconnected from the network, and that take advantage of increasingly sophisticated hardware devices, such as GPS, graphics processing units (GPU) and cameras. Other users need to do advanced ad hoc work with local data and CPU-intensive processing — a great fit for desktop apps. And then, there are users who rely on a continuously connected environment and need focused apps that can be configured for their audience. App revolution is changing every industry It allows professionals to move beyond their desktops and enables many more people to leverage the critical information inside their GIS. This revolution is also obviously changing the way software is licensed. With the new licensing, many of the apps can be shared across the organization and beyond, extending the reach of Web GIS to millions of people without additional licensing. Having everyone always connected to the same authoritative data enables smooth coordination between the field and the office, avoiding the version control issues that can arise with paper-based workflows. Ubiquitous availability and ease of use Challenges traditional licensing system People like ‘apps’ because they are available anywhere and because they provide ease of use for specific tasks. In our industry, the days of generic one-size-fits all GIS applications are over. Industry players that can offer flexible deployment options and a high usability on focused tasks will thrive. Our approach to this is flexible software components for quickly building dedicated apps, and an agile approach in line with Gartner’s “API economy” philosophy. For instance, we see a strong movement toward cross-platform Web apps, exploiting HTML5 and WebGL technologies. But Web is not the only deployment option. Microsoft has invested in getting Windows 10 on mobile devices such as the Microsoft Surface 3 tablet. This brings desktop applications, like LuciadLightspeed, to the mobile market, for example, for use in Electronic Flight Bags (EFBs). We believe this will be a big game changer for our industry. For very high-end applications, virtualization and GPU in the Cloud offer interesting deployment options. The expectation of ubiquitous availability on any device challenges traditional licensing. New business models pop up every day. The only possible answer by our industry is flexibility to adapt to those business models. For instance, users increasingly expect free apps and are only willing to pay for premium services that offer more capabilities, or better data sources. We see the industry moving toward Cloud-hosted apps with free and subscription-based licensing schemes. The landscape is scattered, and apps should really be interoperable. That is why we invest so much in standardization via participation in OGC and other organizations. Frank Suykens, Chief Technology Officer, Luciad Geospatial World • January 2016 • 53 Tech Trends | Apps New and innovative applications for mobile devices Mobile apps are being designed with consideration for the demands and constraints of the devices in use; thus, they are in a way focused and customized solution for every particular use case in each specific user space. Geospatial mobile apps on a smartphone or tablet can provide a way to bring together spatial and non-spatial business information to provide a geospatial context without needing an additional, specialized device. As smartphones and tablets are trending, most people rely on mobile applications for day-to-day work. From your favorite weather tracker or news app to booking a commercial ride, developers all around the world are working hard to develop new and innovative applications for mobile devices, and geospatial is no different. Our smartphones today leverage location to provide us with contextual information based on where we are and what we are doing. In order to provide and deliver a dynamic information experience, it is necessary to fuse four key ingredients — geospatial content, business, and industry workflows and analytics — that not only include geo-processing and spatial analysis, but the visual communication of that information and a design experience for delivering a tailored information to the customers. Simple tools to solve complex problems Traditionally, geospatial companies used to build complex software with the belief that people can use them to answer simple logical things. This has changed enormously and now organizations are more focused on developing simple tools and techniques for complex problems. The new-generation smart apps are cost effective and allow you to use the power of traditional software that can be accessible across the geography, but for different local datasets, making it more customized, agile, focused and user-friendly. Atanu Sinha, Director, India and SAARC, Hexagon Geospatial All the Buzz about Geospatial Biz geospatialmedia Watch the important news, listen to top influencers, get updates from major events 54 • Geospatial World • January 2016 Tech Trends | Automated Cars Cars will no longer be owned, but called up when required T he DARPA Grand Challenge on autonomous cars held on November 3, 2007 saw three winners. The challenge was to navigate an urban and off-road environment of 60 miles, obeying all traffic regulations and taking into account traffic in the form of other autonomous and human-driven vehicles, while achieving a speed of up to 30 miles per hour and completing the course in less than six hours. The DARPA challenge has spawned several commercial ventures. The most visible ones are the 20-plus Google cars which have together clocked nearly a million miles and each are now routinely clocking 10,000 miles a week, mostly in city-driving, which is the average American driver’s annual run. There are four levels of automation. Level 1 is Cruise control which keeps the car’s speed steady at a preset level and is available on most cars today. Level 2 is Adaptive Cruise control where the car’s speed can be automatically varied depending on the speed of the car head, and other features that make the driving experience safe. This too is available in high-end cars. Level 3 is 99% automation with auto-steering. This is the target of most autonomous cars. Level 4 is where the car operates on its own totally without human intervention. Lux Research credits the market for Levels1 and 2 are now about $2.4 billion but is expected to rise to $102 billion by 2030. Levels 3 may be partially achieved by 2030, creating a market of about 56 • Geospatial World • January 2016 $73 billion. But, the study feels that 2030 is too early for full automation because of regulatory issues and lack of reliable prototypes which have make the cut to production. While autonomous cars will result is major savings in terms of reduced accidents and optimized fuel usage, there is a possibility that Level 4 autonomy will actually change the owner-operator model. Cars will no longer be owned, but called up when required. Such a model may result in loss of jobs — but that is another story. Automated driving is well underway It is true that the legal framework is not fully in place yet and that safety concerns need to be addressed as well. We are engaged in active discussions with all our partners, governments and other stakeholders to move things forward in these areas. But generally, good progress is being made. Today, majority of leading carmakers and Tier 1 suppliers are working on technology for self-driving cars. We consider ourselves a partner for these companies in this process, providing a crucial element to make self-driving cars a reality with our highly precise and real-time HD Live Map. Cars have an incredibly rich array of sensors generating tons of data, which currently just sits in the car. The next few years will see us begin to harness that data in the Cloud and do useful things with it. When combining that data with other data from road infrastructure, like connected traffic lights, we can begin to cut journey times and congestion by enabling vehicles to be more aware of their environment. The map is a way of making sense of all this data. Common data format is necessary If we, as an industry, are going to pool data from different vehicles, that data needs to be intelligible and useable regardless of brand. That’s why we, and many others in the industry, have been discussing a common format for the data that gets ingested by a Cloud. We do not have a timeframe, but there is definitely a broad base of support for this initiative, as is in everyone’s interest. There is a general awareness among carmakers that they need to work together as an industry. Bruno Bourguet, Senior Vice President Sales & Business Development, HERE Mainstreaming of autonomous tech Removing the driver from the equation is the single biggest paradigm shift the automotive industry experienced since its inception. However, we think that the horizon to see truly autonomous cars in significant numbers of the streets goes beyond five years. Concretely, in the next five years, we expect more and more ADAS features to become available. We think the mix of vehicles with ADAS or ADAS-like features is likely to increase. We expect to see these features trickling down from high-end models to more mainstream cars and brands, and we expect to see the price tag for these features to decrease steadily over time. Technology to evolve faster than legalities We expect a gradual increase in the penetration of features paving the way to autonomous driving. And it will be no different for autonomous driving. Due to the magnitude of this paradigm shift, we expect the legal environment to catch up with the technology. But, this will happen when the industry can prove that autonomous driving saves lives, reduces emissions and increases the relevance of the automobile product. With the recent takeover of HERE by German carmakers, TomTom remains the only independent, business model-free global map maker for the automotive industry. Our main target users are Tier 1 suppliers and car manufacturers who need to solve the positioning requirements. Jan-Maarten de Vries, VP Automotive Product Strategy & Marketing, TomTom Geospatial World • January 2016 • 57 Tech Trends | Automated Cars Chris Urmson, Director of the Self-Driving Car Project, Google Road safety is top concern About 33,000 people die on America’s roads every year. That is why, so much of the enthusiasm for self-driving cars has focused on their potential to reduce accident rates. One of the most important things we need to understand in order to judge our cars’ safety performance is “baseline” accident activity on typical suburban streets. Quite simply, because many incidents never make it into official statistics, we need to find out how often we can expect to get hit by other drivers. Even when our software and sensors can detect a sticky situation and take action earlier and faster than an alert human driver, sometimes we won’t be able to overcome the realities of speed and distance; sometimes we’ll get hit just waiting for a light to change. And that’s an important context for communities with self-driving cars on their streets. Although we wish we could avoid all accidents, some will be unavoidable. Deeper integration between industries The automotive industry has, in fact, already entered very seriously into the automated driving challenge. Suppliers have developed the necessary technologies in order for higher levels of automation to be researched. And all manufacturers, while already introducing advanced driving assistance systems, are now testing systems to enable higher automation levels. The industry will change toward a deeper integration between suppliers and manufacturers, and will also implement more cooperation with other sectors: electronics, IT, telecoms. We see automation as a progressive process, where we have to solve challenges step-by-step. Several decades will be needed to reach the level of high automation, where the driver is not anymore expected to take control and the vehicle is able to manage any situation. The ERTRAC road map foresee the SAE level 4 of automation to reach the market around 2025. The most common accidents our cars are likely to experience in a typical day-to-day street driving — light damage, no injuries — aren’t well understood because they’re not reported to police. Yet according to National Highway Traffic Safety Administration (NHTSA) data, these incidents account for 55% of all crashes. It’s hard to know what’s really going on out on the streets unless you’re doing miles and miles of driving every day. And that’s exactly what we’ve been doing with our fleet of 20-plus self-driving vehicles and team of safety drivers, who’ve driven 1.7 million miles (manually and autonomously combined). The cars have self-driven nearly a million of those miles, and we’re now averaging around 10,000 self-driven miles a week (a bit less than atypical American driver logs in a year), mostly on city streets. (Excerpts from: The View from the Front Seat of the Google Self-Driving Car) The real push came from automotive industry There is a perception by the media that non-automotive companies are very active and advanced, but this is due to the high media attention that these companies are able to create. In fact, within the automotive industry, huge investments in R&D have been done and are being done on automated driving. Currently, it is the first area of R&D spending by the industry. The push is, therefore, also very much from inside the car industry. Safety is the first priority, and all systems deployed on the market will need to demonstrate extensively their safety. Public authorities are aware about the challenge, and cooperation with the industry has started in order for the regulatory frameworks to be reviewed and adapted. Xavier Aertsens, Director, European Road Transport Research Advisory Council 58 • Geospatial World • January 2016 A Global Communications and Information Company MDA DELIVERS A WORLD OF GEOSPATIAL INFORMATION IN NEAR REAL-TIME MDA is a leading provider of geospatial information products and services. We support users around the world with the complex information products they need for: • Maritime domain awareness • Natural resource and agriculture monitoring • Rapid assessment and response services for disaster events • Near real-time sea ice monitoring • Onshore and offshore oil and gas solutions to help meet environmental, safety and operational requirements MDA operates RADARSAT-2, the high-volume, high-resolution synthetic aperture radar satellite and distributes its radar image data and derived products worldwide. Our growing suite of value-added products and cost-effective services are critical components of efficient operations and informed decision making processes. www.mdacorporation.com Tech Trends | Smart Cities Citizen-friendly solutions the key, not technology fixes T he word smart implies many things, a neat and well-dressed person or a quick-witted and intelligent person. It could also mean a sharp and stinging pain! So what is a smart city? A neatly laid out and ‘intelligent’ city or one that is a pain to live in? Assuming that our city planners do not mean to subject us to further pain, and also assuming that they do not want to raze down a city and build a neat one instead, that leaves us with ‘intelligent’. What makes a city intelligent? A city is an ecosystem built by its inhabitants. It is the intelligence of the inhabitants that makes a city functional, liveable and lively. However, most ‘smart city’ planners tend to think in terms of technology, rapid transit systems, e-Governance, geospatial enablement, optical fibre, routers and switches. Millions are spent and then the planners sit back and find that it does not give the desired results. Technology fixes do not make for intelligence; citizen-friendly solutions are the key. For this to happen the citizens must become a part of the process and not just the passive end beneficiaries of technological ‘smartness’. The UK Department of Business, Innovation and Skills considers smart cities a process rather than as a static outcome, in which increased citizen engagement, hard infrastructure, social capital and digital technologies make cities more livable, resilient and better able to respond to challenges. This can be achieved by the effective integration of physical, digital and human systems in the built environment to deliver sustainable, prosperous and inclusive future of its citizens. A smart city should have free data flow between citizens, government, enterprises and institutions. Such data needs to have a geospatial component because most often the questions are related to location and 60 • Geospatial World • January 2016 routing. Smartness will be decided by the prudent use of its resources, enabled and assisted by geospatial and ICT, to address factors like employment, transportation, environment, safety, low carbon footprint, affordable housing, clean drinking water, electricity, and so forth to create a sustainable lifestyle for its current and future citizens. Michael Dixon, General Manager, Global Smarter Cities, IBM Prospects are limitless the Global ‘Smart City’ Market until 2025 report. A smart city is any urban area that exploits information to optimize the delivery of city services. We can now monitor, measure and manage nearly any physical system at work in our cities. We have the ability to collect and analyze real-time information on everything, from transportation networks to hospitals to the electricity grid. The uses for this information are nearly limitless. It can be used to empower citizens, build political capital, or develop new business models and partnerships with the private sector. It can be used to model and predict how changes to one system will affect others, decreasing the risks of change and speeding the return on investment. And it can be used to draw the businesses that attract talent, and the talent that attracts businesses. Also, “smart cities are anticipated to create huge business opportunities across different industries with a total market value of $1.565 trillion by 2020,” according to Frost and Sullivan’s Strategic Opportunity Analysis of Growth chart of the city data A key element of the smart city is the growing deployment of intelligent devices across the urban landscape, including, for example, smart meters, intelligent building controls, traffic sensors, smart street-lighting, and, of course, millions of smartphones. Together, these devices provide an unprecedented amount of data on the city and its inhabitants. The ability to harness real-time, highly granular data across a wide range of city operations and services is changing the way the urban environment is managed and experienced. The prospects for how this data may be used are enticing for any city leader. Predictive analysis of traffic and transportation patterns can reduce congestion and improve the efficiency of public transportation services. City resources for public safety, social care, and other key services can be targeted more effectively using up-to-date analysis. Energy efficiency programs can be directed at the most vulnerable households and at suitable buildings for retrofit programs. Bring together right resources at the right time The purpose of smart systems is to embed technology into the way the world already functions. We can operate in large cities by using sensors to extract information about traffic flows and utility systems, like water and energy. Analyzing patterns and trends then allows us to make predictions. For electricity, for example, we can predict the usage in a given time-frame so that we can appropriately manage power production. For instance, IBM worked with Stockholm, Sweden, to launch a host of smarter city initiatives to reduce traffic and pollution, by aligning road demand and supply. By helping cities bring together the right resources at exactly the right time, there is tremendous opportunity to provide better service to citizens and make better use of limited funds. Increased access to city data can increase citizen engagement and encourage new forms of creativity and innovation among developers and other service providers. Visualizing the city The greater availability of data and the growing complexity of city management are putting a new onus on the importance of spatial analysis, modelling, and visualization tools. Engineers, urban planners, and property developers, for example, are looking up to modelling tools and GIS-based systems for a long time. Today, new applications are emerging that focus on environmental impacts, mobility strategies, and resilience planning. Eric Woods, Research Director, Smart Cities, Navigant Research Geospatial World • January 2016 • 61 Tech Trends | Smart Cities The people make a smart city ‘smart’ A smart city becomes smart when the data collected can be used among the people, government and private enterprises. It is the people who make a smart city ‘smart’. A smart city goes beyond its sensors, fibre optics and infrastructure. A smart city is where public administrators, private sector and citizens understand the potential of geospatial information in solving key issues and are capable of interpreting geospatial data and formulating questions. Making geoinformation available and accessible is only one side of the coin. The other side is nurturing and developing ‘spatially-enabled citizens’ who are conversant with geospatial information and empowering them to contribute their own visions of a smart city. The role of private sector would be to process this data and provide it for the public use. With a deeper talent pool of ‘spatially enabled citizens’, cities can come up with even more diverse and innovative solutions for their problems. Governments will need to work in tandem with academic institutions, indus- try and other stakeholders to develop capacity in geospatial information and technology, and nurture a new generation of ‘spatially enabled citizens’ for smart cities of the future. Need for strong policy, legal environment Data flow has to be multi-directional, between individuals, businesses, government agencies, and across cities and countries. Hence, we will need to create a strong policy and legal environment, and put in place data policies and standards. Vincent Hoong, Former Chief Executive, Singapore Land Authority Europe and US are taking the lead Most of the current smart city processes originate from Europe or the US. The processes in these regions already have ready access to the spatial and textual data for modelling, computations or to display components of their systems. Highly detailed terrain models are available ‘off-the-shelf’ to assist processes dependent upon gravity or ground shape while the city models are available to define the urban environment. These can be commissioned as well with relative ease owing to the relatively low pricing and abundance of competent suppliers, the proximity of suitable resources, and the efficiency of modern technology. Authorities want, and need to, better manage and utilize their urban environment, and are calling on a range of professional skills to create the tools to deploy smart city concepts. The tools, sensors, services and processes are quite well-developed, and are currently being implemented in the US and Europe. Dearth of suitable spatial data in APAC Unfortunately, in Asia-Pacific, the availability of suitable data is extremely limited. The last decade saw lean demand for such data, hence the drivers to collect and archive the data did not exist. Government legislation often limits the collection and distribution of “off-theshelf” data. Data collection must be bespoke and must be backed by a government agency, so to say. There is no reason why these processes cannot be rolled out to improve the management of cities in Asia-Pacific, except for one reason. The “missing ingredient” is the dearth of suitable spatial and textural data upon which these clever processes can operate. There is a range of technology now available across Asia-Pacific which can help fill that void. David Jonas, Business Development Manager, AAM Group 62 • Geospatial World • January 2016 HELPING YOU MAKE YOUR PROJECT SUCCESS! Switch to KOMPSAT Explore larger area for the same cost KOMPSAT-3 (70cm) Other Satellite (70cm) Other Satellite (50cm) * The graph is visualized based on the list prices. KOMPSAT Cost Effective Client Support Large Coverage Contact us find out more + 82-42-341-0401 www.si-imaging.com sales@si-imaging.com Tech Trends | BIM + Geospatial BIM + Geospatial = Environmentally sustainable cities & infrastructure T oday the building industry is about much more than just bricks and mortar. Not only do you have to build right, but, as Terry Bennet of Autodesk puts it, is it the right thing to build? BIM helps the designers and contractors to scientifically design structures with all facilities coming in to place at the right time, avoiding costly reworks. Is that all? A building is not just a building, but a part of an environment which already exists or is planned. This environment consists of access methods, facilities, utilities and, above all, people who will live and work in them. Just as the building occupies a physical space, so does the immediate environment. This realization has brought about an integration of Building Information Management (BIM) with geospatial to enable the designers to view their creation in the context of the environment in which it is to be built. A marriage of BIM and geospatial results in environmentally sustainable cities and infrastructure. Efficient use of limited resources, better control of schedules and strict adherence to standards are some of the other benefits that can accrue from a cooperation between BIM and geospatial systems. There are some existing standards for BIM sector, such as, IFC from BuildingSMART for open data exchange, CityGML From Open Geospatial Consortium (OGC) for modelling building at various levels and InDoorGML, also from OGC for modelling and sharing indoor space data. InDoorGML is the latest OGC standard focused on mobile indoor location applications, specifically navigation. BIM is expected to increasingly use time 64 • Geospatial World • January 2016 information in the form of 4D data from remote sensing and other sources to monitor the progress of work, detect possible environmental issues and conflict with other activities. The key is to develop intelligent systems that will ultimately lead to smart cities and add to the quality of life of citizens. Terry D Bennett, Senior Industry Strategist, Civil Infrastructure, Autodesk Help in planning efficient infrastructure As the demand for all infrastructure rises globally, the pressure to focus on the right way to increase infrastructure — by taking into account costs over the lifecycle of an asset, and helping to future-proof that asset for growth-increases. All infrastructure is a system and the information infrastructure that supports it is no different. The system requires the capability for monitoring and measuring the degree of change, then analyzing options for addressing it and then communicating those options and their intended impact to drive human actions or in some cases take action without human interaction. This all requires up-to-date accurate geospatial data as the input into a BIM process to supply the context models required to make these decisions. In the longer term, this decision-making process of BIM+Geospatial will help in the planning of newer more efficient versions The future for BIM+Geospatial is bright It has been reported that up to 82% of all smart city initiatives today are thought to have come up short of expectations, and up to 42% are considered to be failures. Due to the failures of smart cities, to merely apply BIM+ Geospatial apps might show some benefits, but will, for the most part, be perceived as delivering less than desirable success stories. As such, it would be far better to apply ‘new’ BIM+Geospatial concepts to programs, such as an intelligent-Enterprise Information Modeling driven ‘Intelligent Cities and Communities Transformation’. We need programs and initiatives that enable ‘smart’ properties, buildings and cities to evolve into ‘intelligent’ properties, buildings and cities. Intelligent Cities 3.0 The use of ‘intelligent’ BIM+Geospatial (combined with CAD and Alpha/Numeric Point Solutions) via i-EIM driven Quad-Directional Connectivity and other applications has enabled the evolution from ‘smart’ properties, building and cities into ‘intelligent’ properties buildings and cities, infrastructure itself. It becomes a positive feedback loop that improves both the physical and digital infrastructure of the city. With the ability to create accurate, intelligent city and infrastructure models in hours and days, it allows us to focus on the problem we are trying to solve, not the data interaction challenges of decades ago. Era of connection As Big Data and near real-time geospatial reality capture (laser scanning, sonar, GPR, drones, etc.) rise, by leveraging this information as an infrastructure type, we are entering a new era — the ‘Era of Connection’. This is where we leverage and extend BIM, combine it with smart systems/sensors and the analytics of geospatial and Big Data to create future ‘smart cities and its supporting infrastructure systems’. In this transition, the role of infrastructure planners and designers changes fundamentally. enabling significant advancements from Intelligent Cities 1.0 to Intelligent Cities 2.0 to Intelligent Cities 3.0. Big Data, which enables ubiquitous access of data, information, knowledge and wisdom, intelligent-enterprise asset management, Intelligent-Building Information Modelling, Intelligent-Geospatial Infrastructure Systems, and Intelligent-Enterprise Resource Planning for everyone, is greatly enhanced with Intelligent BIM+Geospatial apps that readily empower ‘intelligent’ individuals and communities. Advanced focus on social, economic, political issues are enabling organizations to dramatically advance intelligent BIM+Geospatial activities. Lack of funding, timely and cost-efficient education/training and the cooperation between smart technology software and service providers are three of the biggest challenges. Bruce K Forbes, President/CEO & founder, ARCHIBUS Geospatial World • January 2016 • 65 Tech Trends | BIM + Geospatial BIM isn’t just for buildings Merge geospatial with other capabilities The focus on BIM is putting a focus on the relevance of the management of information that geospatial has been doing for a long time. BIM is giving geospatial a stamp of relevance, authority and reason that we have struggled to have before, as well as providing more formal, cross-industry standards. One big factor is the recent emphasis and recognition that BIM isn’t just for buildings, it is for infrastructure. By default, BIM exists within the natural environment and legacy assets and geography; so it requires location and geospatial, to be taken into account. The industry’s move toward 3D/4D and predictive and visualization analytics also fits well with this trend. One big trend is the Internet of things, people, places… and pretty much everything! There is also an increase in using UAVs to do surveys. Satellite imagery is also so accessible now and is offering really exciting prospects for real-time monitoring. Drones and satellites are rapidly transforming surveying and monitoring. Other trends include crowd sourcing, and the way people are now able to move through infrastructure in a more controlled and informed way. This means we need to consider not only the build and operation of infrastructure, but also when it is ‘in use’. This means thinking about how we use robotics, like CAVs (connected autonomous vehicles), moving people and equipment to construction sites or taking our post instead of the postman. Cyber security is a challenge — it can be both an inhibitor and an excuse for not moving to digital data. We will need to get out of our silos and look at other areas like systems engineering. The biggest challenge though is getting people to collaborate. Geospatial can only be fantastic if merged with other capabilities. Anne Kemp, Director (BIM Strategy & Implementation), Atkins Direct impact on asset performance BIM can be defined as advancing the reach and benefits of digital engineering models from design into construction, and from project delivery into asset performance. We have long recognized that BIM is the best, if not the only, way to arrive at better performing infrastructure assets and better performing infrastructure projects. Our cumulative BIM advancements over the past 30 years began with a steadily increasing depth of information modelling (specifically design modelling and analytical modelling) on one axis, and increasing breadth of information mobility on the other. What is recent are our BIM advancements in asset performance modelling. These advancements enable BIM to have a direct impact on asset performance, benefiting owner-operators. Governments must take the lead Governments around the world, such as the UK, the Netherlands, Singapore, and Dubai, are in the process of mandating BIM processes and deliverables for publicly funded infrastructure projects to improve project cost effectiveness and improve the performance of the assets built. Bentley is working closely with a number of these government organizations, including the UK, to help them achieve their goals. BIM and the creation of smart cities go hand in hand. As a result, Bentley’s many years devoted to BIM advancements have resulted in significant contributions to smart city design, construction, and operations. Our products provide for the integrated engineering management. They bring together data from multiple sources. Ted Lamboo, Senior Vice President, Bentley Systems 66 • Geospatial World • January 2016 Policy Pulse | UN-GGIM: The way forward Integrating Ge spatial Information into Sustainable Development Goals T Greg Scott UN-GGIM Prof Abbas Rajabifard University of Melbourne he year 2015 has proven to be a pivotal year for global development. The Millennium Development Goals (MDGs) came to an end, and a new and ambitious development agenda was adopted by world leaders. ‘Transforming our World: The 2030 Agenda for Sustainable Development’ is the global policy to guide the way we collectively manage and transform the social, economic and environmental dimensions of humanity and our planet over the next 15 years. The 2030 Agenda will be anchored by 17 Sustainable Development Goals (SDGs), 169 targets and a global indicator framework, in order to measure and monitor progress. Data, as the basis for evidence-based decision-making and 68 • Geospatial World • January 2016 accountability, will be crucial to the success of the 2030 Agenda. As a geospatial community, we immediately recognize that location and geography are linked to many, if not all, elements of the SDGs. But many have asked, what are the expectations and implications for geospatial information, earth observations, and other location-based information? The answer – considerable! After many years of our combined endeavors to adequately articulate the role, need and value of geospatial information technologies and services to governments and decision-makers around the world, now is the moment in time where we can, and must, elevate and demonstrate our ‘geospatial value proposition’. The global geospatial community, particularly through national geospatial information agencies, has a unique opportunity to integrate geospatial information into the global development agenda in a more holistic and sustainable manner, specifically in measuring and monitoring the targets and indicators of the SDGs. However, the opportunity brings with it substantial expectation to deliver! The 2030 Agenda The 2030 Agenda specifically demands the need for new data acquisition and integration approaches to improve the availability, quality, timeliness and disaggregation of data to support the implementation of the new development agenda at all levels, including to “exploit the contribution to be made by a wide range of data, including earth observations and geospatial information, while ensuring national ownership in supporting and tracking progress.” This need has a goal, target and date associated with it, as described in Goal 17 in data, monitoring and accountability: “By 2020, enhance capacity-building support to developing countries, including for least developed countries and small island developing States, to increase significantly the availability of high-quality, timely and reliable data disaggregated by income, gender, age, race, ethnicity, migratory status, disability, geographic location and other characteristics relevant in national contexts.” In terms of practical action what does this mean for geospatial practitioners in many countries? Firstly, we must consider how these global goals will be measured and monitored, and how progress is made within and across countries. Then, where is the data and evidence coming from? What are the signals or indicators of progress being made and what is the framework and context? The key question is how can geospatial information be implemented and integrated, at a policy level, in order to contribute more holistically to measuring and monitoring the targets and indicators of the SDGs. Not easy when we consider that very little is understood regarding the role of geography in sustainable development processes at the inter-governmental level, including how geospatial information can be applied to sustainable development, and how policies can be implemented to bring the two together in a coherent and integrated manner. Goals, targets, indicators and data The 17 SDGs comprise the aspirational and universal global goals to be achieved by 2030 while the 169 targets provide the detailed objectives to measure progress over the 15 years. But in terms of a robust and annual follow-up and review mechanism for the implementation of the 2030 Agenda, it will actually be the global indicator framework where the data acquisition, integration and disaggregation will be needed. The indicators are the means by which national governments can consistently monitor achievements on, and report 2030 Agenda will be anchored by: 17 1 Sustainable Global Development Goals indicator framework 169 Targets progress toward each of the 169 targets. So the real data inputs and subsequent outputs will be through the indicators, and based on national circumstances. The United Nations Statistical Commission, also responsible for global reporting on the MDGs, established the Inter-agency Expert Group on Sustainable Development Goal Indicators (IAEGSDGs) to develop the global indicator framework, which will be presented for adoption by the Statistical Commission at its 47th session in March 2016. While the development of the indicator framework has largely been a statistical data approach, based on the similar history of the MDGs, the need for ‘geographic location’ in a new era of data needs is well recognized. The statistical community understands that geospatial information and earth observations are able to provide new and consistent data sources and methodologies to integrate multiple ‘location-based’ variables to support and inform official statistics and the indicators for the SDGs. These methods are able to fill data gaps and/or improve the temporal and spatial resolutions of data, by bringing together information from various sources, particularly those related to the environment. Therefore, the The question is how can geo information be implemented and integrated at a policy level to holistically measure and monitor the targets of SDGs United Nations Committee of Experts on Global Geospatial Information Management (UN-GGIM) is working closely with the statistical community, at a national and global level, to provide inputs into the processes to develop the global indicator framework with the IAEG-SDGs. Led by Denmark, a task team of 15 countries has been established to provide these geospatial inputs. Further, at its recent Ministerial Summit, the Group on Earth Observations (GEO) launched an initiative to expand its partnership with UNGGIM to build processes, mechanisms and capacity to integrate earth observations with geospatial and statistical information to improve the measuring, monitoring and achievement of the SDGs. These new efforts are now being combined to determine the geospatial 69 • Geospatial World • January 2016 Policy Pulse | UN-GGIM: The way forward aspects and inputs for the indicator framework. Integrating all information Implementing the SDGs, and measuring and monitoring their progress, will require new and large amounts of data, more rigorous modelling and analysis, and much better data management. It will also take transformative change and collaborative approaches to link different data to geographic location. The location element of data, including statistical data, is going to be critical in the future. As a minimum, ‘sustainable data for sustainable development’ will require consistent statistical, geospatial and earth observation data as the core inputs into the global indicator framework. Figure 1 visualizes what we need to consider in terms of the key areas of data contribution and integration to the indicator process, and informed by good science, enabling technologies, and sound policy. As indicated above, the good news is that these three communities are now very well aligned and positioned to start doing the heavy lifting, and in fact, have already started. Geospatial framework When considering a conceptual integrative framework, the starting point for many countries is the national spatial data infrastructure (NSDI) and the provision of reliable fundamental geospatial data. Although sometimes maligned or misunderstood, the NSDI — or subsequent evolution thereof — is going to be more SD4SD Sustainable data for Sustainable development Earth Observations Geospatial Information Informed by science, technology and policy Figure 1: Key areas of data contribution to the global indicator framework 70 • Geospatial World • January 2016 important than ever, but it must be viewed in a different manner. The NSDI, like the data it provides, must no longer just be seen as a mechanism for mapping and data delivery, but for data integration, analysis, modelling, aggregation, fusion, communication and delivery. The NSDI of the future, fuelled by consistent and best available fundamental geospatial data, can provide the means to organize and deliver core geographies from everywhere (within governments and society) for many national outcomes, including sustainable development. Within the environment of the NSDI, integrating and re-purposing fundamental data into specific thematic topics is not technically difficult and has multiple benefits in many areas, assuming that the data exists, is standards-based, interoperable, and is available. At its fifth session in August 2015, UNGGIM agreed that there is an urgent need for a set of global fundamental geospatial data themes (authoritative, harmonized and homogeneous framework data) to be able to measure, monitor and manage sustainable development processes in a consistent way over time, and to work jointly toward a minimum set of 15 Countries have formed a team, led by Denmark, to provide geospatial inputs fundamental data themes. These efforts are now being initiated and led by UN-GGIM Europe. While this work will evolve in the coming period, we can abstractly consider the high-level framework in which the NSDI could provide the national to global sustainable development geospatial dashboard. As depicted in Figure 2, applying an extended and elaborated set of national fundamental data themes to the 17 SDGs and targets by means of the global indicator framework presents both considerable challenges and opportunities for the global geospatial community. Maximizing the value of fundamental geospatial information to capture elements of the 2030 Agenda, for decision and policy-making, is going to be critical to the future development path of many national geospatial information agencies in the coming 5-10 years. For many countries, especially developing countries, sustainable development will provide a tangible political Data Over Space and Time FUNDAMENTAL Geodetic Elevation Water/Ocean Land Use/Cover Transport Cadastre Population Infrastructure Settlements Admin. Bdys Imagery Geology/Soils Observations Etc. SOCIAL Society Poverty Education Health Population Employment Water Sanitation Equality Gender Governance ECONOMIC Well-being Cities Water Energy Infrastructure Industry Sanitation Economy ENVIRONMENT Water Seas/oceans Land use/cover Ecosystems Forests Agriculture Climate Biodiversity Natural Hazards Pollution Figure 2: Extending national fundamental data themes to the 17 SDGs and targets by means of the global indicator framework ‘trigger’ to accelerate the development and adoption of legal, technical, geospatial and statistical standards; openness and exchange of data and metadata; interoperability of data and information systems; and integration of statistical and geospatial information, including its management and exchange. In other words, the 2030 Agenda will be a trigger to accelerate the development and/or expansion of NSDI strategies. It will also provide a strong consensus on the need to integrate the NSDI within national government’s development plans. An NSDI strategy that is anchored to sustainable development, as an overarching theme, would provide an ‘information’ approach to national policy and implementation. It would also bring the analysis and evidence-base to the process, and thereby a consistent monitoring and reporting framework, that would benefit all areas of government. The expectation is that by 2020 we will be able to significantly increase the availability of high-quality, timely and reliable national data that is disaggregated by a number of characteristics, including geospatial information. This will require collective global leadership, and it will require appropriate frameworks and methods. National geospatial information agencies will need to: collaborate more closely with national statistical and earth observation professional communities; be more unified with similar national to global objectives and aspirations; be delivering consistent and reliable data that is fit-for-purpose; and demonstrate the functionality and value of the geospatial data by integrating it into the wider sustainable development policy process. For its part, UN-GGIM, including its five regional committees, is focusing its activities on achieving such integration and providing a consistent and influential policy voice at the global level. Greg Scott is Inter-Regional Advisor for Global Geospatial Information Management in the United Nations Statistics Division, Department of Economic and Social Affairs Professor Abbas Rajabifard is Head of the Department of Infrastructure Engineering and Associate Dean (International), The School of Engineering at the University of Melbourne. He is also Director of the Centre for Spatial Data Infrastructures and Land Administration, The University of Melbourne 71 • Geospatial World • January 2016 Policy Pulse The benefits from open data are immense G overnments and public authorities across the world are launching Open Data initiatives. Authorities have finally begun to realize the value that opening up data can have on economies. It is also believed that open data policies encourage the use and uptake of geospatial data. The United States, in general, has some fairly broad open data-sharing policies. And yet, for the Landsat data, the first satellite of which was launched in 1972, the United States sold that data for some 34 years or even longer. The data was sold for $500 a scene, when the government was operating the satellites, and for $5,000 a scene when the private sector was operating the satellites. Even then, when the industry technologically got to the point when data can be distributed over the Web, there really were no incremental costs for delivering that data. Finally, in 2008, the USGS and the Department of the Interior announced that all Landsat data will be broadly and openly available. At the peak of data sales in 2001, 51 or 52 scenes a day of Landsat data were sold. The revenue generated from that was barely $4.5-5 million. And even then, it was just government money being re-circulated, because the buyers comprised of federal agencies and universities who received their funding from the government and contractors who were working for federal agencies. And let’s not forget the administrative costs that were being incurred with each transaction. But, soon as the data was released broadly and openly, the orders jumped up to 5,700 scenes a day across the globe. According to an analysis in 2011, the economic benefits of broad open data policy brought back $1.7 billion to the US Barbara Ryan Secretariat Director, Group on Earth Observations 72 • Geospatial World • January 2016 economy and $400 million to other economies around the world. That far exceeds the $4.5 million one federal agency was earlier getting from data sales. More transparency across the world The gigantic success of this decision put a lot of pressure on the rest of the world, with Europe adopting a broad, open data policy for its Copernicus program. I am convinced that Europe will see the same kind of economic returns and hope that these broad open data policies will be adopted elsewhere in the world. With the private sector coming in and building value-added products and services, there will be more transparency, which in turn will help national economies. The benefits are just immense. If most government agencies were to adopt open data policy, it will totally transform our understanding of how the earth behaves as a system. It is not just terrestrial data; it is oceanic data, atmospheric data… And if you really want to start looking at how the earth’s climate is changing, you want data across all those disciplines. We just have not been able to get good global snapshots when we had a business model for data sales, such as, the dust storms off the Sahara are affecting coral reef health in the Caribbean, air pollution coming from Asia, etc. We have to recognize the fact that we are just as connected in the physical world as we are in the digital world. I do realize that we would like there to be big paradigm shifts, but when you think about long-lasting change, sometimes, all you need are incremental steps. Gentle persuasion relentlessly applied is what you need to stay in the game. Integration issues are there We are encouraging the governments to release for public use all earth observation data that they collect — be it from satellites, balloons, UAVs, stream gauges, rain gauges, et al. Our argument is that if taxpayers and governments are funding the agencies building the earth observation systems, the outgrowths of these systems should be made broadly and openly available to those taxpayers. So, we are largely targeting government agencies with earth observation to encourage the adoption of broad open data policies and practices. There is still a lot of public data that we have challenges integrating with. This is also a technological issue because we have not quite got to the point where data standards are the same and data can If taxpayers are funding the agencies that are building the earth observation systems, the outgrowths of these systems should be made available to taxpayers be easily brought in from one sensor to another. On the policy side, making that data available depends on the will of the organization and/or the governments to share their information more broadly. When we start collecting more data on the economic returns and when the governments see their economies benefit from broad open data policies, that will start to speed things up again. So many of our political leaders are concerned about jobs and economic growth, and this is a wonderful, enabling technology to deliver that. Crowdsourced or authoritative data I cannot say that we are there yet with the real ingestion of crowdsourced data yet, but it absolutely will play a big role in the future. I know there are some purists who are very intent on authoritative data, and there are times when that is actually essential. But for many of the other issues, when those authoritative data sources just do not exist, there is a very significant need for crowdsourced information. I would like to cite an example from when I was working with the US Geological Survey. Shortly after the World Wide Web came into existence, the part of the organization that looked at geologic hazards sent out a short message following an earthquake: “Did you feel it?” They asked for inputs like did the books fall off the shelf, did you felt a little dizzy, etc. Then based on that crowdsourced information, the scientists in the organization could see part of the landscape that they had no idea — the sediments were a little different, there were faults in the area that they didn’t predict. There is a wonderful world of crowdsourced information that can actually inform scientific analysis. 73 • Geospatial World • January 2016 Policy Pulse Data privacy is a burning issue E xplosion in data and data analytics has led to increased concerns about data privacy. Common people mostly rely on mobile data, maps and various applications. But then there are uses of geospatial data that are not visible to people. As geospatial data becomes more and more embedded in applications, the real policy question that comes up is when does the rights of the consumer to privacy and to controlling information about themselves are infringed. Insurance companies certainly want more data about you, healthcare companies want more data about you. The balance lies in what you control about yourself and what they know about you. US vs Europe in technological and market landscape mean that those are increasingly under stress. Europeans have taken a more cautious and conservative approach to protecting individual privacy. And it is not an industry to industry question; but really about whom does the consumer trust. Do they trust the companies to protect and manage the data or do they trust and rely upon governments to do that? Or do they trust neither and seek to encrypt more of that data? Innovation in private space Roughly 85 cents of every dollar that NASA had went out to the industry. So it’s not like a lot of it was kept inside. The industry has always been the primary performer of these activities. But what we would always do is making sure that there was in-house intellectual capacity so that government could oversee a good contract manager. And the way to do that is not by reading papers and reports; you had to have people who knew how to build satellites. The scientific community within NASA, for example, always made sure there is at least one or maybe two missions that were being built in-house. So there was a balance between what the governments can do, what their capacities are and where are the sources of innovation, which are normally in the private sector. A classic example of this is that we have weather satellite systems and there have been debates on whether they should be privatized. In general, the In the US, we don’t regulate the data and the information. One of the things that has happened is a realization that the regulatory structures that we have arethe ones that were created in the early Director, Space Policy Institute, 1990s for technologies and Elliott School of International architectures, worked in the Affairs, George Washington University, US early days, but the changes Scott Pace 74 • Geospatial World • January 2016 White House and Congress is the opinion that those satellites are so important for public safety that we do not really privatize them. However, we do see that they are new sources of private data for occultation measurements, and can provide accuracy to weather predictions. There are private firms that are now ready to step out with small satellites to provide this kind of data on a commercial basis so that the government doesn’t have to outlay that funding on its own. Data is wonderful, but we need the craft to command it T impact on someone else’s he National future decisions. This is also Geospatial the reason why NGA has put Intelligence our information on GitHub, Agency has and our apps are available been stressing on iTunes and Google Play. on unclassified geoint for All this not only makes us quite some time now. This more accessible and transmovement has arrived as a parent, but also sparks the part of the technical revoluinnovation that is going to tion. As we put sensors on be the engineering tunnel to everything and put smart where we want to go. devices in everybody’s hands, we need to figure out how to make sense of all this Managing disasters better information. Geography and Putting data in the public geospatial reference are a domain also helps in better great way to do that. Which administration and disaster is why our vision has been management. Take the Ebola pretty consistent: Know the crisis, for example. When Earth, Show the Way and the US President asked Understand the World. We for all hands on deck, we are focused on having such realized that we could help. a command on geospatial in- We recognized that our contelligence that we can enable tent, maps, elevation data, better decision-making. The population data, etc., could bits and bytes do not matter be really useful to a set of if they cannot be used by people wearing white coats somebody in the time frame who didn’t have access to they need it in. We need to classified systems. But, these hone the craft of commanding geospatial intelligence. The information we have Deputy Director, needs to be presented in National Geospatial a way that it can have an Intelligence Agency, US people did have access to the World Wide Web. It took us a couple of weeks to get all the information for them, and I can say with conviction that our data helped a great deal in the understanding of the outbreak. Strategies for growth Data is wonderful, but, it is the people who help to act on that data. So, we really invest in our men and women. Nowadays, we don’t have to work as hard to acquire the data as we used to earlier; the data just comes to us. What we have to work on is to gain such expertise that makes us the best in our profession. Another exciting strategy for growth is partnerships. Sue Gordon 75 • Geospatial World • January 2016 Policy Pulse Spatial data has the potential to drive decision making W hen we started out initially, there was nothing called as spatial database; it was only relational databases of rows and columns. We designed a topologically structured spatial database in the early 1980s. We worked collaboratively with the US Geological Survey and put together a nationwide set of street data and hydrographic data using many different boundaries for collection, tabulation and dissemination of data. We offered that data to communities, states and counties. Our TIGER data is one layer [TIGER products are spatial extracts from the Census Bureau’s MAF/ TIGER database, containing features such as roads, railroads, rivers, as well as legal and statistical geographic areas]. And that is because we topologically enter that information into the file and we maintain it as we add new addresses, make changes, add a new road. We are counting people, we are counting economic activity, we are counting locations, residences. In the early days, we would make that data available on tapes. After the 1990 Census, we put the data on CD-ROM and made that available to people. We want to make the information pervasive. Collaborate and share information Tim Trainor Chief, Geography Division, US Census Bureau 76 • Geospatial World • January 2016 As an active member of the federal geographic data community, we collaborate and share information with all those who use or produce geospatial data. Part of that effort comes together in the form of geospatial platform, which is being able to put all that data in one place. We aim for maximum participation because Census data has the potential to drive several decisions. If you know how many schoolage children are there in a community, you can prepare for the number of schools you need. And if you see a downturn in that, you can again adjust that to ensure that you are not spending more money on something you need elsewhere because your senior population may be growing instead of your younger population. Data visualization makes comprehension easier The US Census Bureau has an active partnership program. In the US, there are about 40,000 functioning governments. We have to understand what their needs are and that we work collaboratively with them to ensure that we are doing as much as we can to support them. That is the benefit of our partnership program. People are realizing that if you have good quality geospatial data and can visualize that data on a map, it is easier to comprehend and see trends. But that only be done if you have the geospatial framework and statistical data that blends together to make that picture come to life. their own solutions but we are focused on connecting networks, connecting business practices, harmonizing policies, using frameworks in a consistent way. Information sharing is a journey, not a destination I nformation sharing and information safeguarding are two sides of the same coin. The things that improve safeguarding also improve sharing. For instance, things like strong and effective modern federated identity and credential access management. A great example of converting the various data feeds into meaningful information would be the National Information Exchange Model. It’s a federated framework which acts as a bridge between different communities, such as, law enforcement, homeland security, intelligence, defense, foreign affairs, etc. It is like a Rosetta Stone, with the ability to bridge concepts across these domains of different activities. Connecting networks, business practises ISE’s [Information Sharing Environment] mission is government-wide terrorism-related information sharing across the entire federal government, with our state and local partners. We knit the information together in a coherent architecture — sometimes at the classified level, sometimes at the censored but unclassified level. Our partners do further analysis to disseminate the information for the frontline law enforcement agencies. Our mission came out of the tragic events of 9/11. We have seen major successes in the last decade, but there is still a lot more to do. We have started initiatives like the nationwide suspicious activity reporting, or the neighborhood watch for the nation. This network has come into play to help in a real-time basis many times, like in the Chattanooga, Tenn., shootings of 2015. We get a lot of demand internationally, asking us to share the best practises of how we share and safeguard Bottom-up and outside-in approach information in the United States. The fact that we can demonstrate and point to specific examples where are aligning national activity with international standard organizations in an open and transparent way generates a lot of buzz and excitement with our international partners. Inherent in what we do is coordination on frameworks in a distributed and decentralized way. Everybody builds The key issue we face is not even about the technology or what the best practices are but about organizational transformation issues. The approach we take is bottoms-up and outside-in. It means working with partners outside the federal government and bringing that voice back into the federal national policy process. We are getting a lot of adoption both on the framework side, as well as with agency partners. Kshemendra Paul Program Manager, Information Sharing Environment, US 77 • Geospatial World • January 2016 Policy Pulse Businesses want geospatial answers to their questions I t is quite a complicated chain of activities that start with change happening in the real world, and our job at Ordnance Survey is to capture the change and make it as real and quick as possible. What we are most particular about is maintaining authoritative quality — how do you signal the quality level of the data that we are presenting to our customers? That is something we are thinking about since this is a source of data that we are not tapping into at the moment. There is no point in having great content and then not liberating it. I think the number of geospatial answers that can be provided is growing exponentially. We want to participate in that and we want to be an innovator and a pioneer. And to do that, we want to work with innovators and pioneers who are also thinking through what the next 10 years are going to bring in terms of change on the planet and how collectively we can help people make sense of that change. Good geospatial infrastructure There have been a number of studies done internally in terms of public investment and public return that show a significant benefit from having a good geospatial infrastructure. On a global level too, there have been a number of studies to find out the impact of a good geospatial infrastructure on a country’s economy. One of the most recent surveys that we participated in found that, in the UK alone, there are businesses worth around £26 billion in annual revenues that are dependent on having a geospatial answer to their questions — insurance to transport or even government delivery. Across the entire planet, there could be businesses worth trillions in revenues that are dependent 78 • Geospatial World • January 2016 on trusted geospatial systems sitting underneath them. Geospatial is a big business and highly relevant to many other businesses, and you quickly get a very big number in terms of the industry which is dependent on us. Another thing which came out of the survey is how innovative these businesses are. So, people doing the survey looked at the percentage of revenue driven by products which have been launched in the last 1-2 years. And the percentage inside the geospatial industry was very similar to the gaming industry and the mobile industry. There are a lot of innovations going on. If you look around the room, you can see that. So, it’s a very exciting industry to be a part of right now. Nigel Clifford CEO, Ordnance Survey, UK Location-based information empowers industries and governments I f I talk about how we enable national security around geospatial technology, we have to first understand that national security is made of economic security, border security and environmental security. And technology underpins everything. Aligning of data allows analysts to explore and understand how the analysis can be differentiated by new data. Technologies come together to give analysts the ability to share data amongst themselves and help them understand how that data can differentiate their answer and empower people who are asking the questions. Users are becoming demanding by the day They are demanding increased content, as well as increased time windows. As we start to run more quickly, the users are asking the systems to work together in order to solve tougher and tougher problems. People have started to recognize what is now in the realm of possibility. This has put new requirements back into the standards processes. The data has to get better in order to feed the system. Government agencies as well as businesses are all dealing with location-based information. And this location-based information is what empowers the industries and gives the government, with support of tax payers, newer capabilities. Location enables quick decision making Whenever there is a human disaster around the world, countries try to respond quickly. And the teams that are responding need to access and assess a lot of information quickly. Much of that information is location-based. And what matters is our ability to provide data as the first part of that response and allow the first responders and the policy makers to get a clear understanding. Whether it is an earthquake in Haiti or Nepal or a volcanic eruption, these responses are important to allow us to quickly ascertain what the problem is and determine a path forward. Our ability to quickly move data and allow people to share their common understanding — whether regarding water purification, tents or medical services — is based on an integrated response coming in. This can only be empowered by geospatial technologies. We believe that having a sharable transparent data allows a shared understanding that provides a response. That said, each government may bring different datasets to the table. Some can easily be shared, some not as easily. But the fabric that is constructed allows for a dialogue to take place and allows us to move forward. Jeffrey Harris Chairman of Board, USGIF & OGC; CEO, JKH Consulting 79 • Geospatial World • January 2016 Policy Pulse There has been a global push for a more relaxed remote sensing data policy C ountries across the world are making satellite imagery data free, and we see this as a very positive development. For long, the Indian Space Research Organization has been trying to convince various government departments about the benefits of space technology. Some of the departments have, in fact, reaped its benefits as well. But, the response from a majority of departments, at best, has been lukewarm. However, now that Government of India is pushing for better utilization of space technology, 60 departments have come on board, and we are trying to meet the demands of their varied programs. This also means that the private industry also needs to get involved. Data policy & security There has been a push for a more relaxed remote sensing data policy. We may blame India for limiting the data release, but one has to be careful with national security issues. I have to agree that making data accessible in certain forms does need to be restricted. The end product should be made available for whatever purpose it is going to be used, but that doesn’t mean that you have to provide the original data in all its forms. This year, we will send a couple of satellites into the orbit to improve the resolution from 0.8 meters to 0.6 meters. We are also working on CARTASAT 3, which will provide a resolution of 0.25 meters. But, that will take at least two or three years more. One of the things we are working on is to make image data available for the users with the help of around 200 servers stationed at various locations. Collaboration is the way All space agencies are currently dealing with fund issues. This is driving everyone to work together and find out common programs. Megha-Tropiques, a collaborative effort between ISRO and the French CNES [Centre National d’Etudes Spatiales], is providing 80 • Geospatial World • January 2016 useful data for weather forecasting. The NISAR [NASA- ISRO Synthetic Aperture Radar project] is another interesting program. One of the key capabilities of this satellite, which will cover the globe once in 12 days, will be repeat pass interferometric observation capability in both L and S bands. NASA is working on the L band portion and ISRO on the S band. The payload integra- A.S. Kiran Kumar Chairman, Indian Space Research Organization, India tion will happen at NASA’s JPL, following which ISRO will launch the satellite. The SAARC satellite concept is also an interesting development in the international cooperation area. We are also working with JAXA and a large number of other space agencies. It is a continuous process and more and more interaction is happening on this. #SpatialChat No such thing as ‘too much information’ Want more? Ask us and our community! @geoworldmedia Product Watch THE WISHLIST As you plan your business moves for 2016, get updated on some of the most promising products and services from top companies Autodesk Autodesk BIM 360 Docs is a comprehensive, Cloudbased service for the construction industry that provides a virtual workspace to create, access, maintain, markup and share 2D and 3D project documents, plans and models. Formerly known as Project Alexandria, Autodesk BIM 360 Docs will help to save time, reduce risk and errors in construction projects. BIM 360 Docs is designed to ensure that the entire project team is building from the correct version of documents and plans. Autodesk BIM 360 Docs connects construction project team members through collaborative, real-time access to construction documents throughout the project lifecycle. The product is expected to come to market in early 2016. Virtual workspace for construction industry Key Features: → Linked 3D and 2D experience, allowing users to interact with models in 2D views and visualize them in 3D on the same page, and vice-versa → Permission-based access control and approval processes to manage the updating and release of documents, preventing project teams from using outof-date information → Blazing-fast viewing experience for large-format PDF design documents, optimized for Apple iOS devices → Automated organization of original and updated construction docs into sets, including highly accurate and customized optical character recognition (OCR) of title blocks 82 • Geospatial World • January 2016 Bentley With ContextCapture, users can quickly produce even the most challenging 3D models of existing conditions for infrastructure projects of all types, derived from simple photographs. Without the need for expensive, specialized equipment, users can create and use highly detailed 3D reality models to provide precise real-world context for design, construction, and operations decisions for use throughout the lifecycle of projects. Easily produce 3D models using photos taken with an ordinary camera, resulting in fine details, sharp edges, and geometric accuracy. Access and share rich reality models of existing conditions for use in any CAD or GIS workflow. ContextCapture will also easily and consistently share information, consumable and accessible, on desktop and mobile devices, in many formats. 3D reality models made easy Key Features: → Unrivaled 3D mesh creation and optimization from ordinary photographs, with high-quality texture mapping and optimal aerotriangulation accuracy → High performance, through multi-core support and optimization for use with general purpose graphics processing units → Extensive scalability through sophisticated tiling mechanisms and “grid computing” support → Robust and flexible publishing capabilities to produce and publish 3D visualization-ready models through a variety of formats and viewer apps with the required level-of-detail – in a way that maintains the highest quality but minimizes hardware requirements Topcon The Topcon rotary-wing unmanned aerial system (UAS) Falcon 8 — powered by Ascending Technologies — is a new mass data collection solution designed for inspection and monitoring, as well as survey and mapping applications. The Topcon Falcon 8 is well suited for smaller mapping projects up to 85 acres and/or construction sites that require high-resolution imaging for documentation, progress monitoring and volume calculations. Rotary-wing UAS expands mass data collection line Key Features: → AscTec®Trinity technology, an autopilot safety feature that provides three levels of redundancy for protection against performance drop or loss of control → Two models available — the GeoEXPERT for surveying, modeling and mapping projects, and Inspection- PRO for industrial inspection and monitoring apps → Easy operation with real-time video and data monitoring capability and navigation software → Photo-tagging and desktop software to quickly generate high-quality material Geospatial World • January 2016 • 83 Product Watch Riegl Riegl’s new VZ-400i 3D Laser Scanner is the evolution of its VZ-400 ultra-versatile terrestrial laser scanner. It combines an innovative new processing architecture, Internet connectivity, and a suite of MEMS sensors with RIEGL’s latest laser scanning engine technology. Real-time data flow is enabled through dual processing platforms: a dedicated processing system for data acquisition, waveform processing and system operations, and a second processing platform which enables real-time data registration, geo-referencing, filtering and analysis to be executed simultaneously. The VZ-400i harnesses this power by streaming it in real-time via the integrated 3G/4G/LTE modem, Wi-Fi, Bluetooth, and Ethernet communications hardware. With its integrated gyroscope, accelerometer, compass and barometer, the scanner’s 1200 kHz pulse repetition rate can be utilized in nearly any environment and orientation. Laser scanning becomes smarter, faster Key Features: → Ultra high speed data acquisition → Real-time registration and processing → Cloud connectivity via Wi-Fi and 4G LTE → 1.2 MHz laser pulse repetition rate → Range up to 800 m; survey-grade accuracy 5 mm Microsoft In 2016, the Microsoft UltraCam team will continue its innovative software roadmap with several new releases of the single, most efficient aerial workflow package UltraMap for generating higher-value data products from aerial imagery. Supported by the world’s first simultaneous nadir and oblique color balancing and de-hazing engine, seamless radiometry will add even more value to users’ data across multi-mission aerial projects. One of the key innovations will be a 3D city engine, capable of generating extremely dense, colorized point clouds as well as textured 3D meshes at unmet geometric quality. UltraMap v5 to come with 3D city engine Key Features: → Create textured 3D meshes at unmet geometric quality → Novel oblique and nadir color balance 84 • Geospatial World • January 2016 → Increased radiometric quality based on the de-hazing module → Flexible radiometric adjustments through interactive user guidance Esri Building GIS apps can be challenging in the face of tight timelines, restricted budgets, and scarcity of GIS developer resources. But, ArcGIS users can easily build native apps that run on any device without any coding, using AppStudio for ArcGIS, an innovative tool from Esri. Any app that you build will run on iOS, Android, Windows, OS X, and Linux devices. And if you are already using ArcGIS, you can convert your maps directly into apps that are ready to use. These native GIS apps can be shared with the public through all popular app stores, or shared within an enterprise. Build GIS apps in a snap Key Features: → AppStudio is designed to enable anyone with GIS skills to configure out-of-the-box templates without coding, so users can leverage existing GIS resources → Whereas GIS departments previously had to build a separate app for each platform they supported, they can now build one app for all platforms and achieve dramatic time and cost savings → Organizations can also customize apps, leverage advanced GIS functionality, or device hardware such as graphics processing units (GPUs), RAM, cameras, microphones, and accelerometers → AppStudio comes with Qt Creator, a cross-platform integrated development environment, and provides the source code for its templates to save developers time Hexagon Safety & Infrastructure Intergraph NetWorks from Hexagon Safety & Infrastructure is a comprehensive GIS data collection and integration solution that leverages Web services to enhance operational insight, customer satisfaction and data quality for utility and communications companies. Key Features: → Realize greater value from their existing GIS data/investment through wider access to data and support of new uses → Improve the currency, accuracy and completeness of enterprise records → Deliver applications that are more user- and task-oriented → Support more agile, affordable and sustainable integration and development Collect and integrate GIS data → Reduce costs in terms of enhanced staff productivity and lower IT costs Geospatial World • January 2016 • 85 Product Watch FARO FARO’s powerful high-speed laser scanner Focus3D X 30 is ideal for a wide range of scanning applications such as architecture, BIM, civil engineering, facility management, forensics and accident reconstruction. The ultra-portable Focus3D X 30 enables fast, straightforward and accurate measurements of interiors, such as small architectural façades, complex structures, production and supply facilities and accident sites. Combining high precision scanning technology with true mobility and ease-of-use, the new device offers reliability, flexibility, and real-time views of recorded data. As the whole range of laser scanners from FARO, the Focus3D X 30 grants highest safety with a class 1 “eye-safe” laser. High-speed laser scanners for interiors Key Features: → Ultra-portable and high-speed scanner → Real-time views of recorded data available → Choice between short-range Focus3D X 30 (30 meters), mid-range Focus3D X 130 and long-range Focus3D X 330 (330 meters) Luciad Luciad has been leveraging the power of GPUs to provide high performance 2D and 3D visualization in its desktop and server product LuciadLightspeed. Now the company has brought the same capabilities to the browser with LuciadRIA V2016. LuciadRIA is a pure WebGL, HTML5 and Javascript solution that allows developers to build beautiful 3D geospatial apps in any modern browser. It enables hardware acceleration using WebGL technology. High performance visualizations in the browser Key Features: → Support for 3D globes and elevation data with detailed terrain visualization → Automatic draping of user imagery and vector data are on the terrain with the same API used for 2D visualization 86 • Geospatial World • January 2016 → Shapes in 3D including points, lines, extruded shapes, etc., → Labeling and dynamic label de-confliction in 3D → 3D Navigation to preview a flight plan or to re-play a recorded flight from the point of view of the pilot → 4D Trajectories visualized and replayed by the thousands Teledyne Optech Teledyne Optech has launched the Galaxy airborne LiDAR sensor with PulseTRAK™ technology. Galaxy achieves high emission rates at high altitude without sacrificing uniform data density across time-of-flight pulse transition zones. Collecting uniform point density in high-relief terrain has plagued LiDAR surveyors for years. With a fixed field of view (FOV), the data swath narrows as it approaches a mountain peak and widens when it approaches a valley. SwathTRAK™ and its in-air dynamic FOV solves this problem by dynamically changing the scan FOV to maintain a fixed-width swath over the ground. Dynamic field of view for airborne LiDAR Key Features: → Equipped with in-air dynamic field of view → Requires fewer flightlines HP → Maintains more consistent point density and XY distribution across the dataset → Significant cost-savings achieved by dramatically reduced collection time PageWide Technology consists of more than 200,000 nozzles on a stationary print bar and spans the width of the page. The portfolio consists of three products: HP’s PageWide XL printer range promises the fastest large-format monochrome and color printing with up to 50% savings in total production costs. These printers allow users to grow with GIS map and point-of-sale (POS) poster printing to generate new revenue streams. HP HP PageWide XL 8000 Printer → Print up to 30 D/A1 pages/minute → Ultra-fast processor, native PDF management → Print on a wide range of media up to 40 inches → Low cost of operation in monochrome and colour versus LED printers HP PageWide XL 5000 Printer → Deliver mixed monochrome and color sets in 50% of the time with a consolidated workflow → Free up the operator — production stacker/online folder, up to 6 rolls, dual ink supplies with auto-switch → HP PageWide XL pigment ink for dark blacks, vivid color and moisture/fade resistance HP PageWide XL 4000 Printer → Up to 8 D/A1 pages/minute monochrome and color → Quick 30-second first page out → Available multifunction device with integrated 1200 dpi scanner or printer For high-quality prints of complex spatial images Geospatial World • January 2016 • 87 Product Watch East View Available in e-book, print and database, Terrain Analysis of Syria and Lebanon contains 47 color topographic maps supplemented with locally sourced reports on demographics, roads, terrain, soils, hydrology, vegetation and climate. It provides simple and easy access to once-classified geographic intelligence covering these entire countries. This publication has been prepared for use by peacekeeping forces, humanitarian organizations, embassy personnel, local authorities and international business investors. Terrain analysis of Syria and Lebanon Key Features: → Maps originallyproduced by Soviet national mapping authorities using on-site surveys, local knowledge and long-term observations → Detailed terrain and geographic reports translated and printed on the reverse sides of 1:200,000-scale topographic maps → Terrain geometry and place names shown in vivid detail Handheld Handheld Group’s ultra-rugged tablets carry high IP ratings and meet MIL-STD-810G standards for withstanding water, dust, shock and extreme temperatures. Nautiz X8 The ergonomic Nautiz X8 is a front-runner in the new → Written analyses offer geographic insights, such as when a river freezes, which off-road areas are traversable in different seasons, and which construction materials are used in local houses → Accessible on computers, tablets and smartphones generation of handheld computers. It delivers the largest, most brilliant capacitive touchscreen in its class, along with an unprecedented combination of processing power, connectivity and field ruggedness. Algiz RT7 The brand-new Algiz RT7 ultra-rugged Android tablet delivers powerful performance for fieldworkers at an excellent value. With a super-fast processor, long battery life, total ruggedness and a host of built-in features, the Algiz RT7 offers heavyweight field performance in a lightweight tablet package. Algiz 10X The powerful Algiz 10X rugged tablet PC can withstand rough weather and harsh handling, and serves as a perfect match for today’s mobile workforce. It is designed for field performance and has the power for critical operations. The Algiz 10X comes with Windows 8 and optional LTE and boasts a quad-core processor. Rugged computers for tough environments TerraGo The TerraGo Edge solution allows organizations to easily collect data and share field information on their smartphones and tablets. It is an open GPS data collection and collaboration platform capable of high-accuracy data collection, seamlessly utilizing any GPS receiver while integrating with any platform including GIS, CAD and other enterprise systems. The latest version (3.8) of the platform features the new TerraGo Edge REST API, auto-recording functions and other enhancements. Collect data and share field information Key Features: → Integrate apps directly with TerraGo Edge Server functionality using simple HTTP methods in JSON format, with any programming language, on any platform, and utilizing Auth 2.0 protocol for authentication → Automatically record GPS lines and polygons (also known as “streaming”) on a mobile device using the TerraGo Edge App, to improve the efficiency of field data collection → Configure TerraGo Edge to record any level of accuracy required, from the on-board location services of iOS and Android devices, to the real-time centimeter-accuracy of Bluetooth-connected GPS receivers RMSI RMSI’s ConflateX is a scalable data conflation solution that addresses the key business challenges of improving spatial accuracy and data integrity of network assets by aligning them to a more precise and accurate real world system. ConflateX has industry specific workflows that can be further customized, utilizing both automated and interactive conflation techniques to achieve operational objectives. The main industries benefitting from this solution include gas distribution, electric, telecom, water and sewer, and transportation. It helps to maintain network data integrity to support business workflows. Other key drivers include optimizes performance of network monitoring systems and improved operational efficiency for emergency and outage management. Solution to optimize spatial investments Key Features: → Integrated workflows to handle large volumes of datasets. Platform independent, supports industry data formats → Conflates network to target landbase, and/or aerial imagery → Business rules driven, highly customizable solution → Programmatically-driven links to ensure higher precision & relative accuracy → Dynamic cell based transformation, to ensure no data distortion → Comprehensive audit trails and in-built quality reporting tools Geospatial World • January 2016 • 89 Product Watch Pix4D Pix4Dmapper software automatically converts images taken by hand, by drone, or by plane, and delivers highly precise, geo-referenced maps, mosaics and 3D models. These products are customizable, timely, and compliment a wide range of applications and software. Users can take images or video with any camera and lens, at any angle. For precise georeferencing, users need to include geotags, image orientation and ground control points. Users will be able to assess the data quality at any stage, even while still on-site. Moreover, automatic generation of georeferenced orthomosaics, DSMs, DTMs, contour lines, 3D point clouds and textured mesh models is also possible. From images to 3D models and mosaics Key Features: → The rayCloud editor combines the 3D point cloud and images used to generate it for a completely new viewing and annotating experience. Assess the quality and improve the accuracy of your results, as well as measure, edit, and automatically classify for GIS and CAD integration DAT/EM Systems To support the increase in UAS data collection, DAT/EM Systems International has come out with Summit UAS (SUAS), a set of tools to critically analyze or extract planimetric and topographic information from UAS data. 3D stereo visualization for UAS → The Mosaic editor empowers you to improve the quality of automatically-generated orthomosaics in just a few clicks. With a real-time local update for faster targeted editing, it’s the easiest seamline editing tool on the market → The Index calculator enables you to create custom vegetation maps, such as NDVI, using multi-spectral imagery with radiometric accuracy. Interpret the maps and produce actionable outputs for direct application in both agricultural machinery and software Summit UAS is built upon the Summit Evolution and LandScape technologies, which include 3D stereo visualization for precise vector feature collection. In 2016, stereo viewing also comes to Summit UAS. The ability to visualize UAS imagery in 3D stereo while drawing vector objects will enable innovative ways to interact with UAS data. The industries for merging 3D Stereo & UAS data processing include farming, forestry, surveying, mineral extraction. conservation and utilities. Key Features: → Uses pre-processed UAS orthophotos and DEM from popular UAV processing systems → Use SUAS tools to generate stereo mate images → View the generated stereo in the SUAS view → Control the cursor in 3D → Collect 3D vector objects → SUAS functionality present in all Summit Evolution editions Leica Captivate See beyond the data Simplify how you work in the field and office with Leica Captivate’s easy-to-use apps and familiar touch technology. Turning complex data into realistic and workable 3D models, you have complete control of any work site. Go further than the data to make the best decisions. Leica Geosystems AG Heebrugg, Switzerland www.leica-geosystems.com Be Captivated Visit www.leica-geosystems.com/becaptivated to find out more and request a demonstration.