badge Tech Siddhi

Monday, 19 March 2018

Philips Lighting introduces LiFi: broadband data through light

Philips Lighting, the world leader in lighting, is now offering Light Fidelity(LiFi), a technology in which high quality LED lighting provides a broadband Internet connection through light waves. As the lighting company for the Internet of Things, Philips Lighting is the first global lighting company to offer LiFi-enabled luminaires from its existing office lighting portfolio.

"LiFi has enormous potential for today's digital age and as the world's leading lighting company we are proud to pioneer new and innovative services for our customers," said Olivia Qiu, Chief Innovation Officer, Philips Lighting.

"While radio frequencies are becoming congested, the visible light spectrum is an untapped resource with a large bandwidth suitable for the stable simultaneous connection of a vast array of Internet of Things devices. Being a lighting company, we ensure that our customers benefit from the finest quality energy efficient light along with state-of-the-art connectivity," she added. 

Broadband Internet connection and quality illumination at the same time
LiFi is a two-way, high-speed wireless technology similar to WiFi but uses light waves instead of radio waves to transmit data. Philips Lighting's office luminaires enabled with LiFi technology provide broadband connection with a speed of 30 Mb per second (Mb/s) without compromising lighting quality. With 30Mb/s a user can stream simultaneously several HD quality videos while having video calls.

Icade, the French real estate investment company, is piloting the ground-breaking technology in its smart office in La Defense, Paris. "LiFi has the potential to be a real game changer in offices. As the leader in our market we wanted to explore the possibilities of this technology for existing and future clients. We plan to showcase the technology in our smart office in La Defense, so aside from stable connectivity, light quality is crucial to us," said Emmanuelle Baboulin, Head of the Commercial Property Investment Division at Icade.  

Benefits of LiFi
LiFi offers benefits over WiFi as it can be used in places where radio frequencies may interfere with equipment, such as in hospitals, or where WiFi signals cannot reach or are weak, such as underground. Other user cases include environments demanding high security; for example, the back office of a financial institution or government service. LiFi adds an extra layer of security as light cannot pass through solid walls and a line-of-sight to the light is needed to access the network. 

How does LiFi work?

Each luminaire is equipped with a built-in modem that modulates the light at speeds imperceptible to the human eye. The light is detected by a LiFi USB key/dongle plugged into the socket of a laptop or tablet (in the future such technology will be built into laptops and devices). The LiFi USB dongle returns data to the luminaire through an infrared link. With Philips LiFi-enabled luminaires, customers get the double benefit of quality, energy-efficient LED light and a highly secure, stable and robust connection as LiFi has 10,000 times the spectrum of WiFi.

NSE and The Garage partnered to host NSE FutureTech 2018 Machine Learning Hackathon.

8 Problem Statements, 30-Hours, 165+ Developers

The National Stock Exchange (NSE) partnered with The Garage, a startup accelerator and co-working space in Mumbai, a business division of JetSynthesys, to host an exciting Machine Learning (M.L.) competition in India - NSE FutureTech 2018 Machine Learning Hackathon. This 30-hour live event held on March 17-18 challenged designers, coders and data scientists from across the country to tackle capital market challenges with machine learning.

The contestants of the event.
Partnering with NSE for the first time, The Garage provided a 170-seater venue and also made available WiFi and beverages for the Hackathon. With pre-registrations of more than 400 developers, the event saw participation from more than 165 creative minds from technological and non-technological backgrounds who worked on real-life problems and attempted to develop a Machine Learning solution within a timeframe of 24-hours.

A panel of distinguished judges including industry experts like Mr. Sankarson Banerjee, CTO, National Stock Exchange (NSE), Mr. Sumeet Vijaywargiya, Associate Director, IBM Watson and Mr. Karan Chaturvedi, Developer Advocate, IBM decided the winners’ basis their skill set and problem-solving ability. The criteria on which individuals and teams were judged was the solution, its technical simplicity, completeness and presentation.

The top three winners of Hackathon received INR 1,00,000, INR. 50,000 and INR. 25,000 in cash, an opportunity to intern at ICICI, complimentary seating at The Garage, along with a bagful of goodies. The winners of NSE FutureTech 2018 Machine Learning Hackathon were The Matrix (Nikhil Lobo, Manish K., Subhnesh Kumar & Prashant), 1st Runner Ups Silicon Alpha (Sanket Shah & Priyank Shah) and 2nd Runner Ups Drupple (Swastik Shrivastava, Anirudh Murali, Rajat Rawat, Shivam Singh & Rohan Chougule).

Speaking about the event, Sankarson Banerjee, CTO, National Stock Exchange (NSE) said, “Machine Learning is making a profound impact in the technology space. It’s diverse ability to learn and adapt to its surroundings allows M.L. to perform at levels similar to or even beyond human capabilities. With Hackathon, our goal was to challenge young minds to solve present day problems in a fun and pragmatic way and provide simple yet meaningful solutions using the latest machine learning tools. We were absolutely thrilled to see the turnout as well as the intelligent minds that are our future!”

Commenting on the same, Rajan Navani, Managing Director and CEO, JetSynthesys, said, “NSE’s FutureTech Machine Learning Hackathon is a great platform for developers to put their abilities to test and showcase their technical knowledge in an exciting M.L. challenge. A one of its kind event, it gave bright young minds the opportunity to learn and compete with the best talent in India. We, at The Garage, are extremely glad to partner with a body like the NSE and foster new talent in the technological space.”

Tuesday, 13 March 2018

Vivo takes HDR to a new level with a dynamic range of 14 EV that adapts to complex lighting

Shenzhen, China, 13th March 2018 – Vivo announced its cutting-edge, AI-powered Super HDR at a media event in China. Following the previous breakthrough success in backlit photography, Vivo is taking mobile photography to new heights with Super HDR's auto scene detection to produce clearer photos, better colors, more details and better tones.
Super HDR follows the same principles as regular HDR but merges more frames to resolve the problems presented by contrasting strong dark and light tones in a photo. Super HDR captures more frames and intelligently merges them to cater to more extreme lighting scenarios.

Great Frames Great Flexibility     
Super HDR's dynamic range can reach up to 14 EV (Exposure Value). With just a single press of the shutter, Super HDR will capture up to 12 frames, significantly more frames than previously possible with existing HDR technology. This provides more image options for merging to perfectly light all the parts of the eventual photo.

Optimized Details Under Extreme Light
To reduce the unnaturally processed look of regular HDR, Super HDR adapts to different scenes and their dynamic range characteristics, selecting and merging frames to best render the image's highlights and shadow details. Super HDR then optimizes shadows to ensure the complex photo's highlights and shadows are balanced and natural.

Get Easy Yet Perfect People Shots
Regular HDR might sacrifice lighting for the person in focus for a better lit photo, but Super HDR ensures the person is clear and well-lit. Super HDR employs smart scene detection to balance the lighting of the person and the background to create a natural feel.

Advanced AI Improves Photos at All Stages
The Super HDR leverages complex AI algorithms to adapt to different scenarios. The moment the shutter is pressed, the AI will detect the scene to determine the ideal exposure strategy and accordingly select the frames for merging. This process ultimately delivers a natural looking shot which looks like it was captured by the human eye.

Advantages of Super HDR

1.      Highly adaptable: With a wider dynamic range of 14EV, Super HDR can ensure perfect shots in a broader range of lighting scenarios, even in adverse ones. This results in image that shows the details of the rock, grass, mountains and clouds despite the back lit, high contrast scenario.

2.      Accurate and smart exposure selection: At up to 12 frames, Super HDR has more lighting options for its smart frame selecting strategies to correctly expose all parts of the photo. This retains the natural contrast that is key to accent the features of this photo.

3.      Intelligent identification: AI-powered scene detection means that each part of the photo can be identified and isolated for processing to insure even complex photos are perfected. The AI can perfectly identify and expose the unique shapes of these rocks, moss, waves and clouds in this shot.

4.      Natural looking results: Super HDR produces more natural tones by reorganizing and optimizing the highlights and shadow portions of the scene when merging the frames and ensures the photo naturally matches the original scene as seen by human eye. This shows the couple clearly and naturally despite sun and umbrella creating such complex lighting.

5.      Perfectly lit portraits: Super HDR prioritizes the best lighting for the person in the photo and adjusts the background exposure to produce an optimal and natural portrait. Optimizing the lighting for the girl and creating complimentary lighting of the sky and flowers help accentuate her in the photo for a perfectly lit portrait.

Alex Feng, Senior Vice President at Vivo added that, "Vivo continues to push the boundaries and provide the ultimate camera experience for consumers. This goes beyond just adding powerful functions, but to developing innovations that our users can immediately enjoy. Today's showcase of Super HDR is an example of our continued commitment to mobile photography, to enable our consumers to shoot professional quality photos at the touch of a button. Using intelligent AI, Super HDR can capture more detail under any conditions, without additional demands on the user."

ad:tech, New Delhi successfully concludes its 8th edition

The 2018 edition of ad:tech New Delhi, the biggest gallery of advertising and marketing community in the country successfully concludes its 8th edition of Digital Marketing & Advertising Conference & Exhibition in India which was held at The Leela Ambience Hotel, Gurugram. ad:tech 2018 delivered two action packed days with more than 90 companies from across the Globe including Google, Adobe, PepsiCo, Future Group, Nestle, Philips, Oath among others who actively participated. Ad tech also witnessed Advertising agencies, publishers and International players from Japan, China, Israel, Russia, US who showcased and discussed the technological innovations in the post advertising era.

The Keynote session on day 2 talked about amalgamation of Automation and Human Experiences Perhaps more than any other industries, marketing has been completely disrupted by technology. Marketers work tirelessly to track customers on their journey across new devices and platforms, and are looking to new and ever-evolving technologies to streamline the process. Prominent keynote speaker Sapna Chadha, Head of Marketing India and Sea, Google said “The biggest challenge today is the pace at which digital marketing is evolving in India is much faster than the architecture in place to support their creativity.” Rashi Goel, Head of Consumer Communications, Nestle shared her views on the topic stating that “The perfect balance between DATA and IDEA is of paramount importance for brands

TNBT winner GetVidAI at adtech 2018

David Shing, VP and Digital Prophet, Oath in his key note session on ‘Innovation is out, Invention is in’ stated that “Today it’s all about packaging! To adopt a new habit we need to break one”. Mr. Shing further added “Creativity should reflect culture. Leveraging data is important. Without data there can’t be creativity and vice versa.”

Elated on the successful closure of 2018 Jaswant Singh, Country MD at Comexposium India said, “We are overwhelmed with the phenomenal response we have received for ad:tech this year. It is important for us to be relevant and to ensure that the platform brings latest trends and innovation for the industry to the forefront.

Jaswant further added, “We are excited to see ourself grow over these 8 years and plan to amplify our presence further in the Indian Market with the launch ad:tech Mumbai in November2018 ”.

Santosh Desai, Managing Director & CEO, Futurebrands in his Keynote session on ‘Towards A Digital Society’ stated that “People experience themselves as individuals on a minute-by-minute basis-liking, clicking, muting. Digital is helping individuals crystallize their sense of self and re-conceptualise their relationship with the world.

The Next Big Thing Challenge

GetVidAI walked away with the coveted ‘The Next Big Thing’ title along with a fully funded pilot project with Nestlé. TNBT Challenge is a global initiative of Nestle in association with Ad Tech with the aim to build a platform that brings entrepreneurs and marketing leaders together and kick start collaborations. This proves to be a great platform for start-ups to prove their mettle by showcasing their indigenous ideas in front of Nestlé and digital media professionals.

Commenting on TNBT Mr. Jaswant Singh said “We are very pleased to provide this platform to budding entrepreneurs. We strongly believe that The Next Big Thing challenge is a great platform for start-ups to push the whole startup Ecosystem.”

The final day saw some of the key and exciting sessions on automation, innovation, integrating AI, Machine learning and building brands the Indian way by renowned brands like Nestle, Google, Oath, Philips, Hero Motor Corp among-st others.
Disclaimer: This post is not edited by Tech Siddhi staff, it is taken from syndicated feed.

Bose Introduces Audio Augmented Reality Platform and glasses to hear

Unveils the Future of Mobile Sound in Eyewear Prototype

This week at SXSW, Bose introduced Bose AR, the world's first audio augmented reality platform, and glasses to hear— a Bose AR prototype that launches the future of mobile sound.  Bose also announces its SDK schedule date for developers, manufacturers, and research institutions, along with collaborations currently under way, and venture funding for related start-ups.

Unlike other augmented reality products and platforms, Bose AR doesn't change what you see, but knows what you're looking at— without an integrated lens or phone camera.  And rather than superimposing visual objects on the real world, Bose AR adds an audible layer of information and experiences, making every day better, easier, more meaningful, and more productive.

Bose AR debuts a minuscule, wafer-thin acoustics package developed specifically for the platform, representing the future of mobile micro-sound with jaw-dropping power and clarity.  It can be seamlessly built into headphones, eyewear, helmets and more, with no compromise to their existing functionality.  And it allows simple head gestures, voice, or a tap on the wearable to control content— replacing the need to swipe, type, or tap a touchscreen for the same commands.

"Bose AR represents a new kind of augmented reality— one that's made for anyone and every day," said John Gordon, vice president of the Consumer Electronics Division at Bose.  "It places audio in your surroundings, not digital images, so you can focus on the amazing world around you— rather than a tiny display.  It knows which way you're facing, and can instantly connect that place and time with endless possibilities for travel, learning, music and more. And it can be added to products and apps we already use and love, removing some of the big obstacles that have kept AR on the sidelines."

The first Bose AR wearablea prototype pair of glasseswas engineered and manufactured by Bose.  They're Bluetooth compatible with microphones for calls, Siri or Google Assistant.  And they debut a new proprietary technology that keeps audio private.  With an ultra-slim, ultra-light, ultra-miniaturized acoustic package embedded discreetly in each arm, they can fit, function and look like standard eyewear, but sound and function more like Bose headphones— delivering amazing, lifelike performance that no one can hear but you.

The Bose AR platform is purposely straightforward, focused on improving daily life instead of extremes.  In addition to delivering stunning audio quality for listening to music, talking on the phone, or using VPAs, a Bose AR wearable uses sensors to track head motion, and the GPS from an iOS or Android device to track location.  The sensors send the motion and location data to a Bose AR-enabled app that aggregates the information, sending relevant, real-time content back to the user's ears instantly.  It's all done hands-free, heads-up, and wirelessly, so there's no need to grab, read or touch the phone. And it can be used for multiple applications.

FOR TRAVEL Bose AR can make every trip easier and more meaningful, transforming sightseeing into sighthearing.  Like simulating historic events at landmarks as you view them— so voices and horses are heard charging in from your left, then passing right in front of you before riding off in the direction of their original route, fading as they go.  Or letting you listen to a renowned speech "pinned" precisely to the famous person in a monument's statue.  Or telling you which way to turn towards your departure gate while checking in at the airport. 

FOR LEARNING Bose AR can make the world a classroom.  Like translating the sign you're reading.  Or telling you the word or phrase for what you're looking at in any language.  Or explaining the story behind the painting you've just approached.

FOR MUSICBose AR can put playlists right in front of you, so you can control them with gestures as easily as voice or touch.  With your wearable on, you can choose or change your music with simple head nods indicating yes, no, or next.

Bose AR doesn't interfere with a wearable's primary purpose— it adds useful information based on where you look.  Like the forecast when you look up, or the restaurants on the street you look down. 

The Bose AR platform is open to approved developers and manufacturers.  Collaborations underway include ASICS Studio, Strava, TripAdvisor, TuneIn, and Yelp.
Bose is also collaborating with academic research institutions, including the MIT Media Lab to advance human interaction related to augmented audio reality, and the NYU Future Reality Lab.
Additional agreements have non-disclosure terms and will be announced in the future.

Bose will invest up to $50 million in start-ups focused on apps, services, or technologies for the Bose AR platform. 

SDK AND AVAILABILITY The Bose AR SDK, and limited quantities of a refined and updated version of the Bose AR glasses, will be available this summer.  Sign-up information and details can be found at