Program

***Please click on session title to view presenation***

Day 1 - Wednesday, October 31, 2018

08:00 - 09:00

Registration

09:00 - 10:35

Welcome

Plenary Hall

Chair: Chen Sagiv, SagivTech 

09:10 - 09:35

Opening: The Challenges of Productizing Autonomous Driving

Amnon Shashua, President & CEO of Mobileye and Senior Vice President, Intel Corporation

09:35 - 09:55

Keynote: AI in the Driver’s Seat - ***VIEW PRESENTATION***

Danny Shapiro, Senior Director of Automotive, NVIDIA

09:55 - 10:15

Keynote: Global Trends and Opportunities in Autonomous Driving ***VIEW PRESENTATION***

Benny Daniel, Vice President - Consulting, Mobility-Europe, Frost & Sullivan

10:15 - 10:35

Keynote: Achieving Mass Commercialization: What will It take for Autonomous Driving to Go Mainstream?

Omer David Keilaf, CEO & Co-Founder, Innoviz Technologies

10:35 - 11:15

Visit the Exhibition

11:15 - 12:15

Sensors

Plenary Hall

Chair: Nate Jaret, Maniv Mobility 

11:35 - 11:55

Detect the Unexpected, the Solid Path to Obstacle Prevention in AD ***VIEW PRESENTATION***

Shmoolik Mangan, Algorithms Development Manager, VayaVision

11:15 - 12:15

Connectivity & Mapping

Hall 2 

Chair: Ran Gazit, Genral Motors 

11:15 - 11:35

Intelligent Vehicle Data Analytics to Put Human in the Loop

Gila Kamhi, Research Lab Group Manager, User Experience Technologies, General Motors

11:35 - 11:55

Using Machine Vision and Machine Learning to Keep Passengers Safe and Comfortable

Guy Raz, CTO, Guardian Optical Technologies

11:55 - 12:15

Drive4U Locate, Affordable Precise and Robust Localization and Mapping for Automated Driving

Paulo Resende, P2/P3 R&D Product Technical Leader , Valeo France

11:15 - 12:15

AI in Autonomous

Hall 3 

Chair: Koby Cohen

11:35 - 11:55

Ushering in the Era of the Self-Healing Car

Zohar Fox, Co-founder and CEO, Aurora Labs

11:55 - 12:15

Next-Gen Computer Vision Cabin Sensing on the Path to Full Autonomy

Inbal Toren, Senior Product Manager, eyeSight Technologies

12:15 - 13:25

Lunch Break

13:25-14:45

Sensors

Plenary Hall

Chair: Prof. Gabby Sarusi, Ben - Gurion University

13:45-14:05

Ride Vision ARAS - 360° Threat Analysis Using Cameras ***VIEW PRESENTATION***

Lior Cohen, CTO & Co-founder, Ride Vision

14:25 - 14:45

One Sensor to Rule Them All- Texas Instruments Imaging Radar ***VIEW PRESENTATION***

Amit Benjamin, Product Manager, Director, Texas Instruments

13:25-14:45

Cyber & Data

Hall 3 

Chair: Gadi Hornstein, Israel Innovation Authority 

13:25 - 13:45

Automotive Data Trends ***VIEW PRESENTATION***

Amir Freund, Chief Product Officer, Otonomo

13:45-14:05

Deep Learning Autonomous Simulation

Danny Atsmon, CEO & Founder, Cognata

14:25 - 14:45

Hacking Automotive Ethernet Cameras ***VIEW PRESENTATION***

Daniel Rezvani , Security Engineer, Argus Cyber Security

14:45-15:05

Securing the Connected Car

Moshe Shlisel , CEO & Co-Founder, GuardKnox

14:45 - 15:15

Visit the Exhibition

15:15-16:35

Closing Plenary 1

Plenary Hall

Chair: David Abraham, Robert Bosch GmbH

15:15 - 15:35

3 Pillars of Autonomous Driving ***VIEW PRESENTATION***

Hilla Tavor, Senior Director, Advanced Development , Mobileye

15:35 - 15:55

Keynote: Autonomous Driving 2.0

Zvi Shiller, Chair, Department of Mechanical Engineering and Mechatronics, Ariel University and the director of the Paslin Laboratory for Robotics and Autonomous Vehicles

15:55 - 16:15

ICT for the Future Mobility

Michael Lipka, Manager Technology Planning, Huawei Technologies

16:15 - 16:35

Keynote: The Role of Safety, Robustness, and Explainability of AI in the Physical World; a Tier-1 Perspective

Mathias Burger, Reinforcement Learning & Planning, Bosch Center for Artificial Intelligence

Day 2 - Thursday, November 1, 2018

08:00 - 09:00

Registration

09:00 - 10:35

Welcome

Plenary Hall

Chair: Rutie Adar, Samsung

09:10 - 09:35

Opening Lecture

Aharon Aharon, CEO, Israel Innovation Authority

09:35 - 09:55

Keynote: Implications of the Software Super Inflation

Oren Betzaleli, General Manager, Harman Israel

09:55 - 10:15

Keynote: The Road to Autonomous Driving: The Challenges and Opportunities of In-Vehicle Connectivity ***VIEW PRESENTATION***

Micha Risling, SVP Marketing and Business Development, Head of the Automotive Business Unit, Valens

10:15 - 10:35

Keynote: Continuous Deep Learning at the Edge ***VIEW PRESENTATION***

Bruno Fernandez-Ruiz, Co-Founder & CTO, Nexar Inc.

10:35 - 11:15

Visit the Exhibition

11:15 - 12:15

Architecture & Simulation

Plenary Hall

Chair: Chen Sagiv, SagivTech

11:15 - 11:35

Autonomous Driving in Unknown Areas

Adham Ghazali, CEO & Co-Founder, Imagry

11:35 - 11:55

How Safe are Autonomous Cars and What is the Industry Doing to Make them Safer – Functional Safety

Ayman Mouallem, Sr. Functional Safety Automation Engineer

11:55 - 12:15

End to End Vehicle Lateral and Longitudinal Control

Xavier Perrotton, Software Department Manager at Driving Assistance Research, Valeo

11:15 - 12:15

Cyber & Data

Hall 3 

Chair: Koby Cohen

11:15 - 11:55

Self Protected Vehicles

Tal Ben David, VP R&D & Co-Founder, Karamba Security

Yoni Kahana, VP Customers, NanoLock

12:15 - 13:45

Lunch Break

13:45 - 14:45

Architecture & Simulation

Plenary Hall

Chair: Koby Cohen

13:45 - 14:05

Mobile Livingroom 2.0 ***VIEW PRESENTATION***

Moritz von Grotthuss, General Site Manager, Gestigon

14:05 - 14:25

On Accidents, Scenarios, Verification and the Path to Safer Autonomous Vehicles

Yoav Hollander, Founder and CTO, Foretellix

13:45 - 14:45

Connectivity & Mapping

Hall 3 

Chair: Micha Risling, Valens 

14:05 - 14:25

Driving the Future of Connected Vehicles ***VIEW PRESENTATION***

Yaniv Sulkes, VP Business Development and Marketing, North America & Europe, Autotalks

14:25 - 14:45

How Deep Learning is Enabling Unprecedented Driver and In-Cabin Monitoring using Cameras

Ophir Herbst, CEO, Jungo Connectivity Ltd.

14:45 - 15:15

Visit the Exhibition

15:15-17:00

Closing Plenary 2

Plenary Hall

Chair: Chen Sagiv, SagivTech 

15:35 - 16:35

Panel- Autonomous Tech – from the Investors Viewpoint

Anat Lea Bonshtien, Panel Moderator, Chairman & Director, Fuel Choices & Smart Mobility Initiative, Prime Minister's Office

Rutie Adar, Head of Samsung Strategy and Innovation Center

Michal Varkat Wolkin, Head of Israel Office, Investments and Innovation, Lear Corporation

Yahal Zilka, Managing Partner, Magma

Danielle Holtz, Director of Business Development, Maniv

16:40 - 17:00

Keynote Lecture

Raj Rajkumar, George Westinghouse Professor of Electrical and Computer Engineering; Director, T-SET University Transportation Center; Director, Real-Time and Multimedia Systems Lab

09:10 - 09:35

Opening: The Challenges of Productizing Autonomous Driving

Amnon Shashua, President & CEO of Mobileye and Senior Vice President, Intel Corporation

The Challenges of Productizing Autonomous Driving

Amnon Shashua

President & CEO of Mobileye and Senior Vice President, Intel Corporation

Prof. Amnon Shashua holds the Sachs chair in computer science at the Hebrew University of Jerusalem. His field of expertise is computer vision and machine learning. Amnon has founded three startups in the computer vision and machine learning fields. In 1995 he founded CogniTens that specializes in the area of industrial metrology and is today a division of the Swedish Corporation Hexagon. In 1999 he cofounded Mobileye with his partner Ziv Aviram. Mobileye develops system-on-chips and computer vision algorithms for driving assistance systems and is developing a platform for autonomous driving to be launched in 2021. Today, approximately 32 million cars rely on Mobileye technology to make their vehicles safer to drive. In August 2014, Mobileye claimed the title for largest Israeli IPO ever, by raising $1B at a market cap of $5.3B. In August 2017, Mobileye became an Intel company in the largest Israeli acquisition deal ever of $15.3B. Today, Prof. Shashua is the President & CEO of Mobileye and a Senior Vice President of Intel Corporation. In 2010 Amnon co-founded OrCam which harnesses computer vision and artificial intelligence to assist people who are visually impaired or blind.

09:35 - 09:55

Keynote: AI in the Driver’s Seat - ***VIEW PRESENTATION***

Danny Shapiro, Senior Director of Automotive, NVIDIA

Danny Shapiro

Senior Director of Automotive, NVIDIA

Danny Shapiro is Senior Director of Automotive at NVIDIA, focusing on artificial intelligence (AI) solutions for the development and deployment of safe self-driving cars, trucks and shuttles. The NVIDIA automotive team is engaged with over 370 car and truck makers, tier 1 suppliers, HD mapping companies, sensor companies and startup companies that are all using the company's DRIVE hardware and software platform for autonomous vehicle development and deployment.

Danny serves on the advisory boards of the Los Angeles Auto Show, the Connected Car Council and Udacity. He holds a Bachelor of Science in electrical engineering and computer science from Princeton University and an MBA from the Haas School of Business at UC Berkeley. Danny lives in Northern California where his home solar system charges his electric, AI self-driving car.

09:55 - 10:15

Keynote: Global Trends and Opportunities in Autonomous Driving ***VIEW PRESENTATION***

Benny Daniel, Vice President - Consulting, Mobility-Europe, Frost & Sullivan

Benny Daniel

Vice President - Consulting, Mobility-Europe, Frost & Sullivan

Benny Daniel, Vice President – Consulting with Frost & Sullivan's Mobility practice, brings with him over 10 years of automotive consulting expertise, with particular expertise covering - R&D Benchmarking; Competitive intelligence and benchmarking; Market Entry and Route to Market Strategy for Glass Manufacturers in Autonomous World; New business model formulation and growth implementation strategy and Pre-due diligence evaluation. Regarded as a domain expert in the electric vehicle market, and automotive R&D benchmarking at a global level, his business model on E-Mobility is globally used by several leading OEMs and utilities.
Benny, a recipient of the Best Consultant of the Year Award for four consecutive years (2009-2012), is known for his ability to understand client requirements and work as an engagement leader and has been the key speaker at several Electric Vehicles and Automotive theme based symposiums.
10:15 - 10:35

Keynote: Achieving Mass Commercialization: What will It take for Autonomous Driving to Go Mainstream?

Omer David Keilaf, CEO & Co-Founder, Innoviz Technologies

Achieving Mass Commercialization: What will it take for autonomous driving to go mainstream?

Fully autonomous vehicles are coming in the not-so-distant future, and it won’t be long before self-driving cars are available to everyone. But what will it take for autonomous driving to go mainstream? What technology is necessary to make it happen? How can the industry make prices reasonable for the masses? What obstacles stand in our way, and how can we overcome them? In this session, Innoviz CEO, Omer Keilaf, will answer these questions as others as he presents a roadmap for achieving mass commercialization of autonomous vehicles.
 
 

Omer David Keilaf

CEO & Co-Founder, Innoviz Technologies

 Omer has spent over 18 years driving cutting edge technologies from inception to commercialization. Before founding Innoviz, Omer led the system and product definition efforts at ConsumerPhysics, building the world’s first handheld molecular sensor for mobile devices. Previous roles include leading the system architecture and engineering teams at bTendo (acquired by ST Micro) and Anobit (acquired by Apple). Omer spent 7 years in the elite technological unit of the Intelligence Corps of the Israel Defense Forces, where he served in leading and critical roles involving large scale technological systems. He holds a BSc and MSc in Electrical Engineering and an MBA, all from Tel Aviv University.
11:15 - 11:35

Staying Ahead of the Curve: Radar and the Future of Autonomous Driving ***VIEW PRESENTATION***

Kobi Marenko, Co-founder & CEO, Arbe Robotics

Staying ahead of the curve: Radar and the future of autonomous driving

The autonomous driving industry requires a sensor that performs at real time in all lighting and weather conditions. In addition, in a world where autonomous cars may drive one towards the other on highways at high speeds, the sensor must be able to “see” them coming from over 300 meters away, track velocity, and detect distance. In this presentation, Kobi Marenko will explain why Imaging Radar is the only technology that can overcome these challenges, and discuss what role Radar will have in an autonomously driven future.

Kobi will focus on how to increase public confidence in the autonomous market, while pushing the industry to develop further. He will specifically discuss the challenges that need to be addressed to achieve a next generation of Radars, such as sensing the road with both an ultra-high resolution and a wide field of view, resolving ambiguities, achieving low false alarm rates, coping with mutual radar interference, while keeping prices low and reliability high.
 

Kobi Marenko

Co-founder & CEO, Arbe Robotics

Kobi is a successful entrepreneur with over 15 years of experience in leading technology and media startups from seed stage to acquisition. Prior to founding Arbe Robotics, Kobi was the Founder and President of Taptica, a mobile DSP acquired by Marimedia, and Founder and CEO of Logia, a mobile content platform acquired by Mandalay Digital.At Arbe Robotics, Kobi and the team work toward making autonomous driving a reality, developing their proprietary imaging radar to provide real-time 4D mapping in high resolution. As the winner of the “Most Innovative ADAS Technology” award from Tech.AD as well as the TechCrunch Disrupt Tel Aviv Contest, Arbe Robotics continues to innovate and be at the top of the real time 4D mapping game.  
11:35 - 11:55

Detect the Unexpected, the Solid Path to Obstacle Prevention in AD ***VIEW PRESENTATION***

Shmoolik Mangan, Algorithms Development Manager, VayaVision

Shmoolik Mangan

Algorithms Development Manager, VayaVision

Dr. Mangan has over 25 years experience in machine learning, machine vision algorithms, physics and optics at Orbotech, Applied Materials and CI Systems. He has a Ph.D. from the Weizmann Institute of Science, as well as MSc and BSc from the Technion Institute of technology.
11:15 - 11:35

Intelligent Vehicle Data Analytics to Put Human in the Loop

Gila Kamhi, Research Lab Group Manager, User Experience Technologies, General Motors

Intelligent Vehicle Data Analytics to Put Human in the Loop

User experience in the era of intelligent vehicles is shifting from the driver to the passenger. We can leverage from in-out cabin sensing/perception abilities of the intelligent vehicles and put human in-the loop and sense/perceive user in his various roles (driver, passenger, road user) in order to automatically adjust vehicle behavior to enhance user experience. In the new world of intelligent vehicles – we first need to address the needs of the  passenger/driver – and target  a natural and personalized in-cabin CAR HMI leveraging from the ability of the vehicle to sense and perceive the cabin and the passengers via inward facing sensors. Then we need to address the needs of the other humans – road users; ensure that new world of mobility  - a world where we humans are surrounded by a growing population of intelligent vehicles is a safer world where  traffic flows - advancing the quality of life.

 

Gila Kamhi

Research Lab Group Manager, User Experience Technologies, General Motors

Gila Kamhi is Research Lab Group Manager, User Experience Technologies at General Motors. She is currently leading a team with multi-disciplinary in-depth expertise covering human-factors, cognitive psychology, big data analytics, artificial intelligence, linguistics, speech and acoustics technologies. The team charter is R&D of user experience technologies covering a wide range of automotive HMI including autonomous vehicles. Prior to General Motors, Gila Kamhi, was a Principal Engineer in Intel Corporation. She led Intel's Perceptual Computing Advanced Technologies team and focused on strategic path-finding as well as perceptually interactive augmented reality experiences. Gila Kamhi holds a MSc degree in computer science from University of Massachusetts at Amherst and a BSc degree from Israel Institute of Technology, She is a co-author of several publications and patents.
11:35 - 11:55

Using Machine Vision and Machine Learning to Keep Passengers Safe and Comfortable

Guy Raz, CTO, Guardian Optical Technologies

Using Machine Vision and Machine Learning to Keep Passengers Safe and Comfortable

As a main goal of vehicles is to transport passengers, the well-being, safety and comfort of the passengers inside a car is of major importance. This becomes especially true when we envision the era of autonomous vehicles, and specifically autonomous public transportation systems such as autonomous taxis, in which there is no driver in the car to assume this responsibility.

At Guardian Optical Technologies we develop an in car multi-sensor to provide rich passenger data. Our sensor combines video images, depth data and micro-scale vibration sensing into one device. Based on these three layers of information we build different applications that can analyze and provide information on different aspects of what’s going on inside the vehicle cabin. The combination of these three different, complementary modalities give us great robustness and reliability.

As an example, a “forgotten infant” application monitors the inside of a locked vehicle to detect if any child (or pet) was left behind. We do this by utilizing the great sensitivity of the micro-scale vibration layer, capable of detecting breathing or even heart-beat motion even with no direct line of sight. Add with the depth and vision layers this result with unmatched detection robustness.

As another example, an “occupancy” detector is used to replace the current seat-belt reminder sensor. We use our sensor not only to detect that there is a person on the seat (and not just a heavy bag), but also to classify their mass and age, and to indicate if they are sitting correctly or are out of position, for smart airbag deployment in case of emergency.

The list of possible applications goes on. It starts with basic passengers and driver monitoring and ends with complex behavior analysis, such as restlessness or violence of passengers. We achieve this level of capabilities by employing state of the art machine learning and neural network algorithms, training the system with big data collection. In this talk we will demonstrate some of our applications and explain how we are able to obtain them.

Guy Raz

CTO, Guardian Optical Technologies

Guy experience is in building and leading multidisciplinary professional teams to solve complex technological challenges. A Ph.D physicists (from the Weizmann Institute of Science) with years of track record in R&D of a wide range of systems and products. Before joining Guardian Optical technologies as CTO, Guy was founder and VPR&D of SiteAware, working on drone based solutions for construction tech. Previously Guy was a researcher at General Motors, working on wide range of automotive sensing technologies, was heading the electrooptics at Pebbles Interfaces (acquired by facebook) and at Elbit Systems lead a few projects and research groups.
11:55 - 12:15

Drive4U Locate, Affordable Precise and Robust Localization and Mapping for Automated Driving

Paulo Resende, P2/P3 R&D Product Technical Leader , Valeo France

Drive4U Locate, affordable precise and robust localization and mapping for Automated Driving

For highly and fully automated driving, a high precise global map is needed to complete and enhance the local perception to handle complex highway and urban use-cases including driving rules and context interpretation. To extract the relevant attributes from the map, decimeter positioning accuracy is required within this map.Commonly used solutions employ high precise GNSS positioning systems like DGPS, RTK which are too costly to be embedded in automotive grade products.To propose a cost effective solution to OEMs, Valeo developed the Drive4U Locate for precise and robust crowd-sourced simultaneous localization and mapping (SLAM). This solution enables centimetric accuracy using onboard Lidar sensors (Valeo ScaLa), affordable automotive grade IMU and low cost GNSS receiver. It works in poor and denied GPS conditions (tunnel, multipath issue...), it is suitable for indoor/outdoor parking and large scale areas and can be complementary to existing HD Maps.For the mapping, multiple vehicles collect and upload to the cloud a real-time online local map based on ScaLa data. A map fusion is done in the cloud to aggregate and update the map changes. This updated map can then be distributed to the subscribed automated vehicles. This allows these vehicles to locate themselves precisely using the updated map.This closed loop process guarantees the maintenance of a precise and up to date map for automated cars.Preliminary results from test drives with Drive4U Locate in urban roads in France and US show that centimetric accuracy can be obtained using current serial production Valeo automotive grade sensors.

Paulo Resende

P2/P3 R&D Product Technical Leader , Valeo France

After obtaining in 2006, a five-year graduate degree in Electrical and Computer Engineering by the University of Coimbra (Portugal), he worked for ESRIN, the European Space Agency center for Earth Observation, in Frascati, south of Rome (Italy).
In early 2008, he returned to Portugal to work for Critical Software at Taveiro and at Lisbon, in the Earth Observation and Command & Control engineering areas.
In mid 2008, Paulo Resende integrated the project team IMARA at the INRIA research Center of Rocquencourt, close to Paris (France), where he particularly worked on trajectory planning and control of driverless and cooperative vehicles (cybercars). 
Since early 2014, he works for Valeo, at the Driving Assistance Research product group, as a System Engineer participating in the development Advanced Driving Assistance Systems (ADAS) for highly and fully automated vehicles, especially on the Valeo's Drive4U project and developing, integrating and demonstrating technologies that will bring highly and autonomous driving into the roads.
11:15 - 11:35

Deep Learning Processing Technologies for Embedded Systems ***VIEW PRESENTATION***

Orr Danon, CEO and VP R&D, Hailo

Deep Learning Processing Technologies for Embedded Systems

Deep learning has become the most prominent technique for image processing in recent years. Alongside its superior accuracy in many tasks and its ability to adopt through training to various use cases, deep learning presents a significant challenge to available processors. The resources required for typical Neural Networks (NNs), when applied to real life use cases such as high resolution, real time sensory data, approach the high 10s terra multiply-and-add operations per second (TMACs), let alone the associated memory bandwidth needed for the task, and pose a significant challenge to existing processors’ technology.
 
Current state of the art solutions for running NNs are better suited for data center environment and are inclined to trade-off efficiency for the sake of higher peak performance. Similarly, latency can be compromised to gain better performance, in which case techniques such as batch processing can be employed. In an embedded environment, power consumption and cost are typically limited by a relatively rigid envelope. Processing must be carried out on the fly, image-by-image. This presents a new set of challenges, which require a new approach for designing processors to best fit the problem domain.
 
In this talk, we will present the problem of image processing in a typical AV/ADAS scenario, both in terms of required resources and of compute paradigm. We will start by exploring the theoretical limits for processing efficiency, compare this with reported results from various players in the industry, and analyze some of the deficiencies which contribute to the large gap between theory and practice for different approaches. A case study analysis will follow. We shall pick a case such as rear-view camera feed for parking assist as an example. This case will be thoroughly analyzed to exemplify some of the statements previously made. This exploration will float some of the inherent system issues, performance requirements and the huge potential unlocked by a capable solution for running NNs locally. We shall conclude with guidelines for designing processing systems which are able to achieve the required efficiency for running state of the art deep learning on embedded devices.
 

Orr Danon

CEO and VP R&D, Hailo

Orr brings more than a decade of experience from the Israel Defense Forces where he led an elite Technological Team to success in some of the largest and most complex interdisciplinary projects in the Intelligence Community. Orr has been recognized with the Israel Defense Award granted by the President of Israel and the Creative Thinking Award bestowed by the Head of Military Intelligence.
Orr holds a B.Sc. in Physics and Mathematics from Hebrew University as part of the Talpiot program and an M.Sc. in Electrical Engineering (cum laude) from Tel-Aviv University.
11:35 - 11:55

Ushering in the Era of the Self-Healing Car

Zohar Fox, Co-founder and CEO, Aurora Labs

Ushering in the Era of the Self-Healing Car

The era of software driven cars is upon us. With all the best intentions, the pressure to deliver car features quicker versus the rise in the amount of software in the vehicle means that not all the bugs can be found and fixed before the features ship. Current diagnostic tools only identify pre-programmed OBD-II errors and current OTA update solutions sit idly by waiting for new software versions to become available. Aurora Labs is ushering in the era of the Self-Healing Car. Using machine learning algorithms to uniquely address all three stages of an automotive software maintenance system, OEMs can now predict, fix, and seamlessly implement OTA updates to faults in the software. With the rising cost and frequency of software-driven recalls, a solution is required that enables reliable and cost-effective rollouts of new automotive features to all ECUs in the vehicle without any downtime for the user.

Zohar Fox

Co-founder and CEO, Aurora Labs

Zohar Fox is the co-founder and CEO at Aurora Labs. He is pioneering Self-Healing Software to make the rollout of software driven functionality more predictable for the automotive OEMs. Prior to co-founding Aurora Labs, Zohar's 20-year career focused on orchestration, design, implementation, and manufacturing of Embedded HW and SW-Cloud based products for TELCOs and technology service providers. Zohar and his teams deployed more than 10M devices in the area of Automotive, banking systems and IOT devices. Over the past several years, Zohar also led product sales in the US and created technology and business alliances with Deutsche Telekom, France Telecom, DSPG, and BOSCH. Zohar was formerly CTO at Rosslare Security products, a 1000-employee access control and home automation IOT technology leader.
11:55 - 12:15

Next-Gen Computer Vision Cabin Sensing on the Path to Full Autonomy

Inbal Toren, Senior Product Manager, eyeSight Technologies

Next-Gen Computer Vision Cabin Sensing on the Path to Full Autonomy

These are exciting times for the automotive industry with the era of autonomous vehicles that is emerging upon us. Drivers now enjoy new levels of safer and better driving experiences thanks to groundbreaking Computer Vision and Artificial Intelligence that is paving way to a driverless future. With all the excitement around autonomous driving it certainly feels that human drivers will soon be a concept of the past, allowing us to rely on calculated machines and eradicate dangerous drunk, drowsy, and distracted drivers, but the reality is that we still have hurdles to pass until we get there. As we pave the path to full autonomy, advanced Computer Vision and AI technologies will play a pivotal role in the transition from low levels of autonomy upwards. The automotive landscape today is dominated by vehicles below level 3 autonomy, requiring an attentive driver behind the wheel, and when looking at road accident statistics reality is grim. In the U.S., auto deaths have jumped tremendously in the last three years with over 40,000 last year according to the NHTSA. Further, they attribute 90% of fatal accidents to human error, while claiming that 74% of car collisions occur due to driver inattentiveness in the 3 seconds prior to the collisions. AAA states that 21% of all fatal accidents involve drowsy driving. Computer Vision and AI solutions are key to battling these statistics by offering driver monitoring systems that provide an ongoing assessment of the driver’s state. Such systems track the driver’s eyelids, blink rate, head pose and even pupil dilatation to detect if the driver is drowsy, distracted or asleep, allowing to take proactive measures from audio alerts and vibrating seats, to reducing the speed or stopping on the shoulder. In level 3 autonomy, where there is transfer of control between car and human, driver monitoring technology accurately assesses the driver’s state to determine the right time for the handoff of control from car back to human.

Inbal Toren

Senior Product Manager, eyeSight Technologies

Inbal serves as eyeSight’s Senior Product Manager, bringing extensive knowledge in strategic product development and user experience gained over 7 years in various domains. As part of her role, Inbal is responsible for defining of the company’s automotive solutions, founding upon deep industry insights to create solutions that comply with the Automotive industry’s standards and needs. Prior to joining eyeSight, Inbal was a Product Marketing Manager at RISCO Group, where she was responsible for the definition and marketing of the company’s full product life cycle in the home security sector. Inbal further brings rich technical experience to her role, having worked at Texas Instruments as a Firmware Engineer in her past. Inbal holds a B.Sc. in Communications Systems Engineering and an MBA from the Honors MBA program from Ben-Gurion University.
13:25-13:45

Future Sensors for Autonomous Driving: Robust Detection of Any Obstacle Under Any Weather and Lighting Conditions ***VIEW PRESENTATION***

Alex Shulman, Director of Products, Foresight

Future sensors for autonomous driving: robust detection of any obstacle under any weather and lighting conditions

 
Self-driving cars are reliant on their sensors to see the world around them. 
One of the biggest challenges standing in the way of autonomous vehicles is the ability to drive under any weather and lighting conditions. 
Poor weather conditions account for 22 percent of automotive accidents in the United States. Weather conditions such as fog, rain, snow can lead to fatal accidents if the driver is not able to detect obstacles and respond appropriately. Autonomous vehicles are designed to provide the highest safety driving levels and must cope with any weather condition to drive safely 
Until today most autonomous vehicle testing has taken place in locations with good weather conditions and focus on the fundamentals of self-driving. However, recently more and more manufacturers have started testing their self-driving cars in adverse weather conditions, as they understand the need for autonomous vehicles to drive safely under heavy rain at night, dense fog or heavy snowfall. Such testing reveals the strengths and weaknesses of different sensor systems.
It is a common understanding that autonomous driving will require data fusion from multiple sensors for redundancy purposes and for increased sensing robustness. For example, dense fog or heavy snowfall create significant challenges for visible light cameras and for LiDARS. Radar on the other hand can better cope with this condition, but it lacks the required resolution to allow safe autonomous driving.
Foresight has developed a unique multispectral vision sensor. It is based on seamless fusion of 4 cameras – 2 sets of stereoscopic long-wave infrared (LWIR) and visible-light cameras, enabling highly accurate and reliable obstacle detection. Simultaneous information from both regular (visible) and thermal (far-infrared) stereo cameras increases dramatically detection of pedestrians, vehicles and other stationary and moving objects under severe weather and lighting conditions.
The goal of this presentation is to assess the accuracy of different sensors in severe weather conditions. Specifically, detection accuracy and detection range are compared using sensor data recorded day and night, simulated fog and heavy rain for different types of obstacles (vehicles, pedestrians, other static and moving objects). Detection accuracy and range for Foresight multispectral vision sensor, is assessed using (a) stereo visible light images; (b) stereo far-infrared images; and (c) combination of visible and far-infrared stereo images.

Alex Shulman

Director of Products, Foresight

Alex Shulman serves as Director of Products in Foresight and he oversees all aspects of product management and product marketing. 
Mr. Shulman has deep experience in multiple high-tech product categories: opto-electronics, mechanics, software & algorithms in several successful startups that became leading Israeli high-tech companies. 
He has led several successful new product launches resulting in significant market share gains in highly competitive markets worldwide. Mr. Shulman holds degree in engineering and MBA.
13:45-14:05

Ride Vision ARAS - 360° Threat Analysis Using Cameras ***VIEW PRESENTATION***

Lior Cohen, CTO & Co-founder, Ride Vision

Ride Vision ARAS - 360° Threat Analysis using cameras ***VIEW PRESENTATION***

In the word where all major cities are congested and the traffic hell grows each year, the two-wheeler vehicles (PTWs) can be the driver for gaining back insanity.  
Air pollution, time saving and money savings are sufficient triggers to promote the two-wheelers. In many metropolitan cities the use of the PTWs has recently grew at a double-digit percentages.
A national organization Ride To Work Inc., encourages PTWs commuting and draws lots of participants and awareness to the benefits of motorbikes. 
 
However, many people perceive the PTWs as dangerous vehicles !!!
Advanced driver-assistance systems (ADAS) are designed to increase car safety and help the driver in the driving process. While ADASs are now very common in cars, in the PTWs space there are almost no solutions that relates to rider assistance systems.
 
Ride Vision develops Advanced Rider-Assistance Systems (ARAS) that provide features like  collision avoidance, blind spot warnings, lean angle warnings, adaptive cruise control, automatic signaling and more... The system is based on two wide-angle cameras and on the algorithms that fuse these sensors to understand the road state and alert the riders on the particular PTWs threats.

Lior Cohen

CTO & Co-founder, Ride Vision

Lior is an avid motorcyclist that harnesses the power of pragmatic thinking, artificial intelligence and computer vision to make motorcycles safer. Before Ride Vision, Lior was a VP R&D at PicScout where he lead a team of visual algorithm researches and engineers. Prior to that Lior lead mobile and homeland security development teams dealing with complex algorithmic problems.
Lior holds a M.Sc. in information system engineering from the Ben-Gurion University.
14:05-14:25

Invisible Light, Invisible Data: Leveraging SWIR to Solve the AV Visibility Challenge

Avi Bakal, CEO & Co-founder, TriEye

Avi Bakal

CEO & Co-founder, TriEye

Avi Bakal is a multidisciplinary leader, and CEO at TriEye. TriEye is a Semiconductor startup that develops a disruptive solution which adequately addresses the challenges of smart and autonomous vehicle safety under the adverse weather/low visibility conditions. Avi is an experienced R&D engineer and large-scale projects manager. Avi has a BSc in computer and electrical engineering and an MSc in applied physics specializing in optics and lasers.
14:25 - 14:45

One Sensor to Rule Them All- Texas Instruments Imaging Radar ***VIEW PRESENTATION***

Amit Benjamin, Product Manager, Director, Texas Instruments

One Sensor To Rule Them All – Texas Instruments Imaging Radar

***VIEW PRESENTATION HERE***

This presentation describes the current ADAS/AV sensing technology market, 
survey the advantages and disadvantages of each sensor type in light of few of the accidents which recently occurred, and then introduce the Imaging Radar sensor, a new sensor technology, that we believe will be a must have sensor in level 3, 4 & 5 AV.
In the 2nd part of the presentation, we will go deeper and explain how cascade multiple TI single chip radars to a high performance radar sensor. The cascade radar system can support both MIMO and TX beam forming modes for high angle resolution and long detection range. The high accuracy phase shifter enables actively beam steering towards desired angle of interest. Some field test results are presented based on TI 4-chip cascade evaluation board to demonstrate the achieved performance

Amit Benjamin

Product Manager, Director, Texas Instruments

Amit Benjamin is a Product manager director from Texas Instruments,  leading the mmWave RFCMOS  Radar sensor product definition and key customers engagements in the Automotive ADAS market. Professional experience also includes roles in system engineering and Sales management
13:25 - 13:45

Automotive Data Trends ***VIEW PRESENTATION***

Amir Freund, Chief Product Officer, Otonomo

Amir Freund

Chief Product Officer, Otonomo

Amir is Otonomo’s Chief Product Officer, responsible for setting the company strategic definition of the product & technology offering. Amir Freund is a highly experienced entrepreneur, previously a co-founder of Alvarion (NASDAQ: ALVR) & a co-founder and CEO of ForNova. During 2015 - 2017 Amir was leading the Ford Motor Israeli activity, involved in Ford acquisitions / investment in Israel and creating collaborations with >40 local automotive startups.  
13:45-14:05

Deep Learning Autonomous Simulation

Danny Atsmon, CEO & Founder, Cognata

Deep Learning Autonomous Simulation

Danny Atsmon

CEO & Founder, Cognata

Danny Atsmon, an expert in ADAS and deep learning, is the CEO at Cognata Ltd., a dynamic young technology company that brings the disruptive power of artificial intelligence, deep learning, and computer vision to simulated testing for autonomous cars. He has been in the business of launching high-tech products for more than 20 years. Before joining Cognata, Danny served as Harman’s (NYSE:HAR now Samsung) Director of ADAS and Senior director of Machine learning. He has co-founded two startup companies, Picitup, a computer visual shopping suite, and iOnRoad, which uses a phone’s native camera and sensors to detect vehicles in front of a car (later acquired by Harman International). Danny holds several United States utility patents and has created a pipeline of dozens of patent-pending applications. He has also won many top industry awards, including Design and Engineering Showcase Award (2012) for innovation in design, CTIA award (2012) for Best Mobile Application for Automotive, Safe Driving & Transportation, Microsoft Think Next (2012), and the QPrize (2013), which is awarded after Qualcomm Venture’s seed investment competition. Danny is a graduate of the prestigious Israeli Defense Forces (IDF) Haman Talpiot program, where he served in the elite Unit 8200, and holds a B.Sc. degree in Physics from Tel-Aviv University.
14:25 - 14:45

Hacking Automotive Ethernet Cameras ***VIEW PRESENTATION***

Daniel Rezvani , Security Engineer, Argus Cyber Security

Hacking Automotive Ethernet Cameras

Hacking Automotive Ethernet Cameras
Daniel Rezvani, Argus Cyber Security

***VIEW PRESENTATION HERE***

Autonomous vehicles rely on a range of sensors to interact with the world around them. One of the most noticeable sensors is the Ethernet camera - a standard camera which is used for vision-based ADAS (advanced driver assistance system). Since this camera has become a critical part of an autonomous car’s safety (it is responsible for identifying nearby hazards, traffic signs, etc.), the consequences of a successfull cyber-attack launched against such a camera can be devastating, resulting in real physical injury or death. In this technical presentation, you will learn how our team of automotive cyber security researchers were able to easily hack an Ethernet camera (similar to those being integrated into today's connected vehicles) and trick it into thinking the pre-recorded video is reality.

 

Daniel Rezvani

Security Engineer, Argus Cyber Security

Daniel Rezvani is a cyber security Engineer at Argus Cyber Security. Prior to joining Argus, Daniel served in the elite IDF cyber intelligence unit (8200) where he led a team of cyber security researchers. Daniel holds a B.A. in Computer Science from the Interdisciplinary Center Herzliya.
14:45-15:05

Securing the Connected Car

Moshe Shlisel , CEO & Co-Founder, GuardKnox

SECURING THE CONNECTED CAR
The automotive industry is on the brink of a technological revolution to develop self-driving, autonomous vehicles. Connected cars have become the stepping stone for autonomous vehicles, resulting in an industry shift of focus, priorities, and resource allocation. No longer are manufacturers or consumers solely concerned about general metrics and performance measures such as fuel efficiency, mileage or vehicle longevity Consumers want personalization and customization with hints of flexibility. They crave
easy access to applications and services to facilitate their hectic days, in their own
respective styles. With increased connectivity comes an exponentially increased risk for malicious hacking attempts albeit ransoms, intent to inflict harm, vehicle disabling, or data theft.
SHIFTING THE PLAYERS IN THE AUTOMOTIVE INDUSTRY
The dynamic automotive connectivity ecosystem includes various stakeholders and players that are not solely restricted to automotive business - creating opportunities for non-automotive industry players like IBM and Intel to create value added customizable applications and services such as temporary engine upgrades or increased tire traction. This aligns with in-vehicle consumer trends, wants, and needs: ease of accessibility to features such as navigation, live traffic updates, access to their mobile device to receive messages and phone calls on the go, and vehicle personalization. Guardknox firmly believes that cybersecurity is the foundational layer for added connectivity and customization . It is essential to employ the proper cyber defense solution for not only privacy and protection, but more importantly, safety of one’s self and their loved ones.
Guardknox approaches cybersecurity by borrowing tactics from the defense sector, as IT
solutions may not be appropriate for moving platforms. GuardKnox was recently awarded for its defense perspective on vehicle cybersecurity framework at Frost & Sullivans 2018 Innovation Awards Banquet.
THE CHALLENGES
Cybersecurity is starting to be monitored and enforced by regulatory bodies and authorities. OEMS and Tier 1 suppliers are collaborating with cybersecurity companies to provide comprehensive solutions. Most current solutions available are essentially IT-based or learning solutions. This creates a slew of challenges to maintain integrity and security during design, implementation, and over the air updates such as: Due to their architecture, these solutions are learning based off of algorithms  Heuristics based solutions generally are around 98% reliable, leaving room for false positives Software only solutions cannot meet safety standards or common criteria GuardKnox is the only solution to adhere to the stringent ISO26262 and Common Criteria to ensure 99.9999% reliability rate. The vehicle is the smallest unit in which we house our families, any room for false positives is unacceptable.
GUARDKNOX AUTOMOTIVE CYBER SECURITY
GuardKnox is an automotive cyber-solutions company providing comprehensive cyber defense.The GuardKnox team brings decades of experience providing the same cyber security solutions to the Israeli Air Force systems: Iron Dome, Arrow III, Israeli F15, F16 and the Israeli F-35. GuardKnox’s patented deterministic hardware and software solution, the SNO ™ , or Secure Network Orchestrator, is completely autonomous, does not require any constant online connectivity and can defend against both known or unknown attacks. The SNO ™ can be seamlessly installed as a complete hardware and software solution during manufacturing or retrofitted in the aftermarket. Due to GuardKnox’s flexibility, the hardware architecture can be independently licensed or the software stack can be implemented as a separate entity. GuardKnox’s patented hardware architecture, supported by the Communication Lockdown ™ methodology, physically separates the different networks of the car, while forcing every message to pass through its three tiered detection system: the routing, content, and contextual layers. Every bit in every field in every message is monitored and locked.
PROTECTION ACROSS INCREASED CONNECTIVITY
GuardKnox’s SNO ™ houses the patented secure service oriented architecture, or SOA. SOA allows unified communication as well as access control and service level partitioning. The SOA components of an ECU reports data about the vehicle to a central system using a separation kernel to abstract and conceal communication across platforms. This enables a compartmentalized approach to ensure each partition is secured separately to facilitate the security demands associated with higher levels of connectivity. In doing so, OEMS and Tier 1s can offer additional revenue generated applications and personalized services under secure protection.
The GuardKnox solution family provides security in-depth with a central gateway ECU drop-in SNO ™ , a local SNO ™ for externally connected ECU's, as well as a reporting mechanism providing centralized fleet security. As a complete ECU, it integrates seamlessly into the vehicle, the value chain and the vehicle production process. Furthermore, the central and local SNO ™ can be retrofitted after production and in the aftermarket as another source of additional revenue streams for OEMs and Tier 1 manufacturers.
CLOUD CONNECTIVITY: PALO ALTO NETWORKS® PARTNERSHIP
In addition to in-vehicle vulnerabilities, the channel between the ECU and the cloud is the basis for communication between the OEM and the updating vehicle functionality. This channel is also a potential security gap through which hackers could gain control of the vehicle. GuardKnox has recently partnered with Palo Alto Networks ® to create an end-to-end solution from in-vehicle to the cloud to the OEMs backend to support predictive maintenance, data analytics and AI, logging services and more by:
● Securing on going communications between OEMs and vehicles
● Securing communications between vehicles and database applications
Palo Alto Networks© GlobalProtect™ Cloud Service secures the external network between the vehicle and the OEM cloud through the use of their secured communication channel while GuardKnox provides a holistic lockdown approach to the vehicle’s internal network by enforcing strict set of rules on all incoming and outgoing communication in real time.

GuardKnox

Moshe Shlisel

CEO & Co-Founder, GuardKnox

Moshe Shlisel serves as GuardKnox Cyber Technoligies CEO since 2015. 
He brings with him over 15 years of successful managerial experience in various global companies, including specific experience in management technology companies marketing, sales and business development. He served as the CEO of Trumedia , COO of SatixFy, a Fabless semiconductor Company where he was responsible for; Business Development, Marketing, Operations & Project Management. Prior to this Moshe served as Executive Vice President & General Manager of ExactCost Israel and as a Senior Director at Elbit systems, a leading provider of advanced, high-performance defense electronic and systems. He retired as a Lieutenant Colonel from the IAF. He currently also serves as a visiting professor at Tel-Aviv University and the Interdisciplinary Center
15:15 - 15:35

3 Pillars of Autonomous Driving ***VIEW PRESENTATION***

Hilla Tavor, Senior Director, Advanced Development , Mobileye

3 Pillars of Autonomous Driving

***VIEW THE PRESENTATION HERE***

Hilla Tavor

Senior Director, Advanced Development , Mobileye

Hilla is Senior Director of Advanced Development at Mobileye. She manages the Advanced Development department overseeing Mobileye's future technology business development activities, that lead all pre-production/advanced development programs. With a Computer Science and Math degree, Hilla has held in the past various software development management positions in several tech companies. Hilla lives in Rehovot with her wife Noga and 4 children.
15:35 - 15:55

Keynote: Autonomous Driving 2.0

Zvi Shiller, Chair, Department of Mechanical Engineering and Mechatronics, Ariel University and the director of the Paslin Laboratory for Robotics and Autonomous Vehicles

Autonomous Driving 2.0

Zvi Shiller

Chair, Department of Mechanical Engineering and Mechatronics, Ariel University and the director of the Paslin Laboratory for Robotics and Autonomous Vehicles

Professor Shiller is the Chair of the Department of Mechanical Engineering and Mechatronics at Ariel University, director of the Paslin Laboratory for Robotics and Autonomous Vehicles, and Chair of the Israeli Robotics Association.   He earned the B.Sc. degree from Tel Aviv University, and the M.Sc. and Sc.D. degrees from MIT, all in Mechanical Engineering.  Before joining Ariel University, he served on the faculty of the Department of Mechanical and Aerospace Engineering at UCLA.  Professor Shiller's research activities have focused on optimal motion planning, obstacle avoidance in static and dynamic environments, navigation of off-road and intelligent road vehicles, and on assistive robotics.
15:55 - 16:15

ICT for the Future Mobility

Michael Lipka, Manager Technology Planning, Huawei Technologies

ICT for the Future Mobility

Individual transportation is hit by three mega trends, which are electrification, automation, and the digital transformation. All these three mega trends are also vital questions for the traditional car manufacturer and therefore of tremendous impact on the automotive eco-system existing for more than 100 years. Electrification simplifies the drive train enabling new player in the automotive market on the one side of competition while automation will place a taxi ride into a fully new competitive position against private car ownership. In consequence car sharing offerings in major economies will be boosted, and in turn sharing customers will be open for new modalities like the electrical vertical take-off and landing (eVTOL) vehicles being on the horizon as a convenient alternative for medium distances. Despite both approaches, privately owned cars from traditional OEMs and any car sharing fleet, will become electric and autonomous, the technological approach might be different. Traditional OEMs are used to offer their customers flexible individual all-in-one vehicles and therefore developing a corresponding stand-alone autonomous driving experience. In contrast car sharing operators are focused on return on invest within a fleet approach. Consequently fleet automation will follow other paradigms in comparison to car automation and will therefore drive other technological solutions. A supporting road and back end infrastructure in case of fleet automation incl. flying vehicles, decoupled from energy consumption constrains of the BEV electronics, drives potentially other solutions as currently developed within car industry. Fleet operation will especially benefit from the digital transformation driving administrative expenditures to a very low end.

A draft scenario of these developments is going to be presented as a discussion base within the community to develop a joint understanding of future technology demand. Will these developments drive competition or will the community see a cooperative environment? We intend to initiate a discussion on technology development for the future automobile eco-system.

Michael Lipka

Manager Technology Planning, Huawei Technologies

Michael Lipka received a diploma in communication engineering from the Technical University of Darmstadt in 1991 and a PhD in semiconductor technologies for RF devices from the University of Ulm in 1996. Different management positions in the communication industry from 1991 to 2007, in particular working for Alcatel, Siemens, and NOKIA with focus on technology management for narrowband switching and mobile communication systems. From 2007 to 2016 project manager for strategic long term technology planning activities within Siemens Corporate Technology. Development of long term visions for different Siemens businesses. Since 2017 director for technology planning in Huawei’s European Research Institute.
16:15 - 16:35

Keynote: The Role of Safety, Robustness, and Explainability of AI in the Physical World; a Tier-1 Perspective

Mathias Burger, Reinforcement Learning & Planning, Bosch Center for Artificial Intelligence

The role of safety, robustness, and explainability of AI in the physical world; a Tier-1 perspective.

Artificial Intelligence (AI) has been a story of great success over the last years in many fields and application domains. Technologies such as machine learning or planning - often grouped under the general term AI – are substantially contributing to make all kind of products more intelligent and human friendly, as we will illustrate on several Bosch products. It is widely accepted that AI technologies are also key to autonomous driving and the Internet of Things. Specifically in recent years, the focus in the automotive space was about perception and in-vehicle UX. However, as AI technologies are increasingly applied in the physical world, ‘under the hood’ and in the production line, new levels of rigor and guarantees are required. We will discuss the role of safety, robustness and explainability of AI, showing where research, which contributes to develop highly reliable standards for AI, is needed in the engineering domain

Mathias Burger

Reinforcement Learning & Planning, Bosch Center for Artificial Intelligence

Mathias Bürger is heading the research group on Reinforcement Learning and Planning at the Bosch Center for Artificial Intelligence. He joined Bosch in 2014 and since then worked in several research activities at the intersection of artificial intelligence and robotics. Currently, he is also principal investigator for the H2020 project Co4Robots. He studied Engineering Cybernetics at the University of Stuttgart and the University of Toronto and holds a PhD degree in control theory from the University of Stuttgart. For his PhD work he received the 2014 EECI PhD Award in recognition of the best PhD thesis in Europe in the field of Control for Complex and Heterogeneous Systems.
09:10 - 09:35

Opening Lecture

Aharon Aharon, CEO, Israel Innovation Authority

Aharon Aharon

CEO, Israel Innovation Authority

Aharon Aharon is the CEO of the Israel Innovation Authority, an independent public entity that operates for the benefit of the Israeli innovation ecosystem and Israeli economy as a whole. Its role is to nurture and develop Israeli innovation resources, while creating and strengthening the infrastructure and framework needed to support the entire knowledge industry. Prior to his role at the Authority, Aharon served as the VP of Hardware Technologies and General Manager of Apple Israel. His previous roles were Co-founder, CEO, and chairman of Camero, a startup developing through-the-wall micro-power radars, active Chairman of the Board at Discretix, a startup providing solutions for data security, CEO of Seabridge, a wholly owned subsidiary of Siemens and COO of Zoran. Prior to that, he held a number of senior management positions at the IBM Research Division. Aharon holds a B.Sc and M.Sc in Computer Engineering and Electrical Engineering respectively from the Technion, where he lectured for more than 15 years.
09:35 - 09:55

Keynote: Implications of the Software Super Inflation

Oren Betzaleli, General Manager, Harman Israel

Implications of the Software Super Inflation

In the past years, cars are becoming more and more dependent on electronics and software. Modern cars have between 50 and 100 Electronic Control Units (ECUs), from smaller micro-controllers to powerful computers running open operating systems such as Android. These cars also have up to 100 million lines of code. As a comparison, this is more than 10 times the number of lines of code in NASA’s space shuttles, and 5 times more than the new Boing 787 Dream Liner airplane. It is estimated that software today accounts for almost 10% of the car value (BOM).
When we look 5 to 10 years ahead, these numbers are expected to grow significantly, with the introduction of ADAS and self-driving features, digital cockpits, cloud services, heads up display and more. Estimates indicate somewhere between double and triple the amount of lines of code, to between 200 and 300 million lines of code, and up to 30% of the BOM.
The implications of this super inflation of software in future cars includes the obvious fear of bugs and ultra-costly recalls that may result, but also implications on design, production, supply chain, service and more. Car makers (OEMs) already started to get themselves prepared, mainly with the introduction of Remote Software Update solution, but they will have to think deeper on several other changes they will have to make to be ready.
In his presentation, Oren will review multiple implications of the software inflation, and the different aspects OEMs need to think about when addressing the issue, and preparing for the implications.
Oren is leading HARMAN’s Remote Software Update business, who was already selected by 20 OEMs, representing almost 300 million connected vehicles in the coming 10 years.
About HARMAN
With a workforce of some 30,000 automotive professionals across the Americas, Europe, and Asia, HARMAN designs and engineers connected products and solutions for automakers, consumers, and enterprises worldwide. HARMAN’s broad-array of cutting-edge products includes digital cockpit systems, Remote Vehicle Updating over-the-air (OTA) solution, audio products, SHIELD Cybersecurity Solution and IoT services. In March 2017, HARMAN became a wholly owned subsidiary of Samsung Electronics Ltd.
Audiophiles from every generation call on HARMAN to deliver the best in sound in the studio and on the stage, in the car, at home and on the go. HARMAN’s portfolio of legendary audio brands includes AKG®, Harman Kardon®, Infinity®, JBL®, Lexicon®, Mark Levinson® and Revel®.
Since 2012, HARMAN has made significant and ongoing investments in Israeli technology, including the acquisitions of iOnRoad, Redbend and TowerSec. HARMAN Automotive Cybersecurity is a full service global business unit that has deep experience in traditional IT security and embedded security, as well as several years of pioneering work in automotive cyber security. HARMAN’s Remote Vehicle Updating offering is the only OTA solution that enables efficient full-vehicle management, already chosen by 20 OEMs to enable secure OTA software updates to more than 30 million vehicles.
In November 2017, HARMAN has announced the launch of the International Cyber Security Smart Mobility Analysis and Research Test (SMART) Range in Israel, in cooperation with Ben-Gurion University of the Negev. HARMAN has been an active member of industry institutions - such as Auto-ISAC, SAE/ISO, Jaspar and others – working to enhance cybersecurity awareness, design and deployment across the global automotive industry.
 

Oren Betzaleli

General Manager, Harman Israel

Oren Betzaleli is the General Manager of HARMAN Israel, and a Senior Vice President of HARMAN International, responsible for the Software Platforms Product Business Unit which includes various Software Products such as Automotive Cloud Platform, Remote Software Management (OTA), Device Management, Automotive Cyber Security, Device Virtualization. 
Oren has joined HARMAN from the Redbend acquisition in 2015, where he was EVP of Products and Marketing. Prior to that Oren was 4 years in Retalix in similar role, before it was acquired by NCR, and 12 years in Amdocs, most of them in senior product management roles.
 
09:55 - 10:15

Keynote: The Road to Autonomous Driving: The Challenges and Opportunities of In-Vehicle Connectivity ***VIEW PRESENTATION***

Micha Risling, SVP Marketing and Business Development, Head of the Automotive Business Unit, Valens

The Road to Autonomous Driving: The Challenges and Opportunities of In-Vehicle Connectivity

The hype and buzz around the autonomous car has been going on for a while, and chances are, it will only increase. The potential benefits that autonomous cars will bring to end users and cities are appealing, and as technology continues to advance, the reality of the autonomous car will solidify.
 
A major element in the road to autonomous driving is the under-the-hood, in-vehicle connectivity infrastructure. As important as the cameras, displays, sensors, and other devices are important for the autonomous car, they must be connected to each other and to the car’s computing devices to enable actual autonomy.
 
This ability to provide effective and efficient communication within the car is the key to the autonomous car. Vehicle manufacturers are now facing the challenges of in-vehicle connectivity:
 
  • The need for increased bandwidth, as more and more cameras, sensors, and telematics are added
  • The need for a reliable, light-weight and easy to use cable for transmitting data
  • The need for robust and adaptable connectivity to handle the rough automotive environment
  • The need for a scalable solution
This session will look at these challenges and what are the implications in meeting (or not) these requirements, while focusing on the opportunities that new technologies are offering, such as extended PCIe connectivity, fast camera links, symmetric and asymmetric transmissions and more.
 

Micha Risling

SVP Marketing and Business Development, Head of the Automotive Business Unit, Valens

With more than 20 years of experience in marketing and R&D executive roles in established NASDAQ and startup companies, Micha heads the company’s automotive business unit and leads the company’s marketing and business development activities in several verticals. Over his tenure at Valens, he played a major role in establishing the company’s reputation as a leader in the digital connectivity space. Micha is also the Chair of the HDBaseT Alliance’s Marketing Committee, where he is responsible for the organization’s strategy and marketing activities.
10:15 - 10:35

Keynote: Continuous Deep Learning at the Edge ***VIEW PRESENTATION***

Bruno Fernandez-Ruiz, Co-Founder & CTO, Nexar Inc.

Continuous Deep Learning at the Edge

***VIEW PRESENTATION HERE***

The robustness of end-to-end driving policy models depends on having access to the largest possible training dataset, exposing the true diversity of the 10 trillion miles that humans drive every year in the real world. However, current approaches are limited to models trained using homogenous data from a small number of vehicles running in controlled environments or in simulation, which fail to perform adequately in real-world dangerous corner cases. Safe driving requires continuously resolving a long tail of those corner cases. The only possible way to train a robust driving policy model is therefore to continuously capture as many of these cases as possible. The capture of driving data is unfortunately constrained by the reduced compute capabilities of the devices running at the edge and the limited network connectivity to the cloud, making the task of building robust end-to-end driving policies very complex.

Bruno Fernandez-Ruiz offers an overview of a network of connected devices deployed at the edge running deep learning models that continuously capture, select, and transfer to the cloud “interesting” monocular camera observations, vehicle motion, and driver actions. The collected data is used to train an end-to-end vehicle driving policy, which also guarantees that the information gain of the learned model is monotonically increasing, effectively becoming progressively more selective of the data captured by the edge devices as it walks down the tail of corner cases.

Bruno Fernandez-Ruiz

Co-Founder & CTO, Nexar Inc.

Bruno Fernandez-Ruiz is cofounder and CTO at Nexar, where he and his team are using large-scale machine learning and machine vision to capture and analyze millions of sensor and camera readings in order to make our roads safer. Previously, Bruno was a senior fellow at Yahoo, where he oversaw the development and delivery of Yahoo’s personalization, ad targeting, and native advertising teams; his prior roles at Yahoo included chief architect for Yahoo’s cloud and platform and chief architect for international. Prior to joining Yahoo, Bruno founded OneSoup (acquired by Synchronica and now part of the Myriad Group) and YamiGo; was an enterprise architect for Fidelity Investments; served as manager in Accenture’s Center for Strategic Research Group, where he cofounded Meridea Financial Services and Accenture’s claim software solutions group. Bruno holds an MSc in operations research and transportation science from MIT, with a focus on intelligent transportation systems.

11:15 - 11:35

Autonomous Driving in Unknown Areas

Adham Ghazali, CEO & Co-Founder, Imagry

Autonomous Driving in Unknown Areas

Autonomous driving has come a long way albeit it's success is limited only to those areas where a High-Definition map is pre-built and known. Inspired by the recent achievements in Artificial Intelligence particularly of methods that combine trees and DNNs (e.g. Alpha Go Zero), this talk demonstrates how to effectively combine Deep Learning and convention path planning to drive in unknown areas.

Adham Ghazali

CEO & Co-Founder, Imagry

Adham Ghazali is the CEO of Imagry. He spent the last 10 years working on various machine learning problems including large scale computer vision, Brain computer interfacing and Bio-Inspired facial recognition. He is interested in the intersection between biology and computer science. At his current post, he is responsible for strategic R&D and Business Development.  
Prior to cofounding Imagry, Adham was a brain researcher focusing on the study of the visual system in infants. 
11:35 - 11:55

How Safe are Autonomous Cars and What is the Industry Doing to Make them Safer – Functional Safety

Ayman Mouallem, Sr. Functional Safety Automation Engineer

How safe are autonomous cars and what are the industry doing to make them safer – Functional Safety

Functional Safety is all about ensuring the safety of operation of electronic devices, in all conditions, even when hardware failure happen. With the rise of autonomous driving, Functional Safety is gaining more and more importance and its becoming key requirements in all automotive electronic systems. In this presentation, we introduce the standard ISO-26262 that is enforced by car OEMs to guarantee the Functional Safety of all devices they put in the car, how the standard is used to communicate requirements between OEMs, Tier-1’s and Tielr-2’s, we touch on its requirement for system-level, software-level, and electronic chips level.

Ayman Mouallem

Sr. Functional Safety Automation Engineer

Optima is the company to resolve the number the #1 challenge faced today by semiconductor companies, that is inhibiting and slowing down their penetrating of the automotive segment of semiconductors: Functional Safety. Autonomous Driving is revolutionizing the structure of the car, converting it into supercomputer-on-wheels, with 10’s of huge and complex chips and systems. These chips need to meet the stringent Functional Safety requirements of ISO-26262 like ASIL-D, with no adequate solution in the market before Optima’s.
11:55 - 12:15

End to End Vehicle Lateral and Longitudinal Control

Xavier Perrotton, Software Department Manager at Driving Assistance Research, Valeo

End to End Vehicle Lateral and Longitudinal Control

In this presentation, we will present a complete study of an end-to-end imitation learning system for lateral and speed control of a real car. Indeed, as autonomous vehicles are accumulating mileage, there are two main research streams to deal with increasingly difficult situations, in particular urban driving, with a high amount of reliability to get increasing levels of autonomy.
In the most common approach, the autonomous driving functionality is cut into different modules, from perception to control, going through data fusion, path planning etc. An alternate method is to consider the vehicle behavior in an end-to-end fashion: the raw low level data from the sensors (classically cameras) is more or less directly used to produce the car commands. This is described as “behavior reflex”, because it is a reaction to the environment without explicit description. Using alternate approaches such as end-to-end, even for limited tasks, can be combined with more classic pipelines, in order to provide new redundancy channels that are mandatory for safety purposes.
To train a neural network to produce end-to-end controls, one can let the network learn by itself the way the vehicle should behave by trial and error: this is Reinforcement Learning. Another way of training would be to take the ground truth from an external behavior (typically, a human driver), and train the network to produce the same control output using the raw signal. In literature, this is called behavioral cloning or imitation learning. In this study, we will describe how to design an end-to-end imitation learning system for Lateral and speed control of a real car, learn two networks, one for lateral control and one for longitudinal.
Concerning networks to control steering angle for autonomous cars, most of research publications use multiple long range cameras to generate failure lateral cases in order to handle error accumulation issue when testing on real car. In this presentation, we present a novel model to generate this data augmentation using only one short range fisheye camera. We then present how we build a simulator and how we used it as a consistent metric for lateral prediction evaluation. Experiments are conducted on a new dataset corresponding to more than 10000 km and 200 hours of open road driving. We evaluate this model on real world driving scenarios, open road and a custom test track with challenging obstacle avoidance and hard turns. The final model was capable of more than 99% autonomy on urban road.
Concerning lateral control, we propose a neural network based on an LSTM and propose to use a loss that integrates dynamic information in order to let the system learn speed control dynamic similar to the one produced by a human driver. We also propose data augmentation and label augmentations that are relevant for imitation learning in longitudinal control context. Based on front camera image only, our system is able to correctly control the speed of a car. More precisely, the system is able to control the speed of a car in simulation environment, and in a real car on a challenging test track. The system also shows promising results in open road context.
Finally, we will describe how we have integrated the two networks on Valeo research prototype cars, used as platforms, both for collecting data and for testing. Cars were equipped with an active drive by-wire control, which made it possible to send steering wheel angle and acceleration commands to the car. To run the network inference on board, an embedded GPU board was used. Our experiments show that the proposed system is able to drive the lateral and control car speed in both simulation and real cars. We will present results of these experiments and show videos of open road test and demonstrations done at CES 2018.

Xavier Perrotton

Software Department Manager at Driving Assistance Research, Valeo

Currently Valeo's Software Department Manager at Driving Assistance Research, Perrotton is leading a department of research engineers on research and innovation projects around autonomous driving.
He develops state of the art machine learning and computer vision methods to power cool and innovative products. Xavier Perrotton's expertise includes computer vision, AI, deep learning, applied math, software architecture, sensors and high performance computing. After working on Augmented Reality, creating prototypes and contributing to their industrialization through project such as MiRA (an augmented reality solution already deployed in Airbus), he is now focusing his research on AI & Autonomous Driving. 
11:15 - 11:55

Self Protected Vehicles

Tal Ben David, VP R&D & Co-Founder, Karamba Security

Yoni Kahana, VP Customers, NanoLock

Tal Ben David

VP R&D & Co-Founder, Karamba Security

Tal is experienced in developing new, innovative products into viable, market-changing products. He was a Senior Research and Development Manager at Check Point Software Technologies (NASDAQ:CHKP), where he oversaw the development of endpoint and mobile security technology. Prior, he gained extensive cybersecurity experience overseeing the development of Check Point’s big data analytics and event correlation products. Before entering the private sector, Tal served as a computer science officer in an elite technological unit with the Israeli Defense Forces (IDF). Tal has a B.Sc. in Computer Science and Statistics from Bar-Ilan University. He is married to Osnat and a father of two children.

Yoni Kahana

VP Customers, NanoLock

Mr. Kahana brings more than 20 years of experience in managing¸ leading and developing large-scale projects in secure telecommunications and embedded systems¸ from idea-stage to completion in R&D¸ product and business environments. Prior to joining NanoLock¸ Mr. Kahana was the Product Cybersecurity Group Manager for General Motors (GM) in Israel. At GM¸ he managed the Israeli Cybersecurity Group responsible for securing crucial elements in the car. He also was a Security System Team leader and Security System Product Engineer at Qualcomm. He served as an officer in the Israel Defence Forces (IDF) in elite unit where he led and developed cybersecurity solutions.
11:55 - 12:15

Challenges in Implementing IoT in Your Products and Building Your Organization Go-To IoT Strategy ***VIEW PRESENTATION***

Rami Khawaly, Co-Founder and CTO, MindoLife

Challenges in Implementing IoT in Your Products and Building Your Organization Go-To IoT Strategy

Customer enthusiasm for the Internet of Things (IoT) is growing. Obviously, IoT is no longer a buzzword or passing phase. However, as more companies seek to offer connected devices, they find that this mission hides a lot of obstacles, know-how, and capabilities to create IoT devices from scratch. In a rush to market, many companies neglect to consider all the security aspects of deploying IoT device, preferring to “develop first, add security later.” The results are IoT projects schedule and budget overruns, security risks, and long-term support difficulties.
 
A lot of our customers say they faced many challenges when developing IoT solutions, such as the right security, integration with current systems, properly operate in the strict limitations that IoT devices allow (e.g., low energy, low computational capabilities) and eventually achieving returns on their investment.
 
What are the main IoT implementation challenges?
 
* Security concerns
There are four main reasons for building security into an IoT network:
1. Malfunction - IoT makes it possible for the predefined hardware to perform a specific task automatically in the real world. 
2. Abuse – Connected devices can be used by external hackers for malicious purposes. 
3. Intrusion – Connected devices can be a gateway for hackers to reach sensitive data from internal enterprise systems.
4. Consumer demand – in a 2017 survey by IDC, 25% of companies surveyed stated that security is the main hindering factor for deploying IoT solutions with their organization.
Keeping your IoT devices safe from hackers hijacking and attacks is an on-going task that starts with a deep cyber security design and will need a continues maintenance, security is an attitude and not a matter of encrypting data.
 
* Unclear business model
Along with every new emerging field, many questions arise regarding the right business model and what is the optimal way for the service provider of making money, this question affects both the supplier and the consumer of the new product.
 
* Lack of standards
Lack of IoT standards becoming a threat. According to Gartner, a total of 8.4 billion people and devices will be connected to the Internet of Things (IoT) this year. Despite the increasing popularity of IoT solutions in companies, 63% of applications are actually linked to the consumer sector.
 
* Low Power Devices Management
 
Most of the IoT devices have low capabilities, whether it has a poor CPU or a minimal RAM or a limited power supply, in many cases all of these weaknesses are combined in one device.
These limitations are common in devices like sensors, detectors, and wearables which must be properly managed to enable all IoT network needed features while taking into consideration the limited resources available.
 
* Mass Production Line Support
 
The IoT abilities of your devices can be an integral part of your mass production process, it’s your decision whether the device will cooperate optimally during the procedure of the production or not!
 
* Continues support & ongoing updates (OTA)
Five questions to ask yourself when considering an OTA update technology in the IoT context:
 
Does the OTA update mechanism support automatic recovery from failed updates while preserving its connection to the update server?
 
Is security a first-class feature of the OTA mechanism and not something bolted on as an afterthought?
 
Can OTA updates be applied in an efficient manner, minimizing resources like network bandwidth, storage, and compute?
 
Can OTA updates be applied at various levels within a given application e.g., application configuration updates versus device firmware updates?
 
Does the OTA update mechanism leverage container technology (or similar) work seamlessly across diverse hardware and software environments?
 

Rami Khawaly

Co-Founder and CTO, MindoLife

Co-Founder and CTO Of MindoLife, Rami is a veteran software engineer with hands-on experience in designing and creating secure IoT networks and solutions. Previously, Rami was an R&D Team Leader at Vigilance Networks and has been consulting in complex development issues to R&D companies since 2011. Before that, Rami held development positions at several software companies including RealCommerce & Clal insurance. Rami holds a B.Sc. in Software and Management Engineering from Afeka Tel Aviv Academic College of Engineering.
13:45 - 14:05

Mobile Livingroom 2.0 ***VIEW PRESENTATION***

Moritz von Grotthuss, General Site Manager, Gestigon

Mobile Livingroom 2.0

Most expect the car to be the third place of living aside from our home and our office. Longer distances to commute, more traffic, more time in the car is creating this reality already today. But how can we avoid that this development turns out to be our personal mobile dystopia? How do we create a place to be and a place where you like to be? HMI, AR, Interior Cocoon and new perceptual safety features are core to create a Mobile Livingroom 2.0. - These concepts will be introduced, explained and discussed.

Moritz von Grotthuss

General Site Manager, Gestigon

Moritz von Grotthuss is a tech-nerd, innovation enthusiast, automobilist and entrepreneur. After some successful (but somehow boring) years in legal, sales and consultancy, he followed his heart and joined gestigon as a late founder and CEO in 2012. Today he is responsible for the strategy of the company, his collection of frequent traveler cards, and most of the commercial stuff. gestigon has been working closely with the automotive industry since more than 5 years and demoed a number of behavior tracking solutions at different shows including CES, IAA and others. Moritz has been driving this strategy and the focus of gestigon to have an automotive DNA since day one. Early 2017 gestigon was acquired by Valeo and is now a center piece of Valeo’s Interior Cocoon strategy to further improve automotive safety and enable enhanced usability and user experience. Additionally to his job as CEO of ‘gestigon – a Valeo brand’ he is part of the technology and startup scouting team of Valeo/corporate. Moritz is married, has kids and loves - if he is not working or at home - outdoor activities.
14:05 - 14:25

On Accidents, Scenarios, Verification and the Path to Safer Autonomous Vehicles

Yoav Hollander, Founder and CTO, Foretellix

Accidents, Scenarios, Verification and the path to safer autonomous vehicles.

That tragic Uber accident has brought Autonomous Vehicle (AV) safety into sharp focus. It raised awareness of aspects like how people are more afraid of things they can’t control, the need for third-party testing, the insurance implications of all this and so on.
Regardless of the specifics of this incident, this presentation will look at the bigger picture of AV deployment, safety and verification, and expand on the following claims:

Beyond a certain safety threshold, AVs should be deployed
While safer than human drivers, AVs will continue to have many fatal accidents
AV manufacturers and regulators should employ a well-thought-out, comprehensive, continuously-improving, multi-execution-platform, transparent verification system

There seem to be a need for the various stakeholders (the public, lawmakers, regulators, AV manufacturers etc.) to agree on some general framework for handling these accidents (and the whole deployment process). That framework should ensure, among other things, that:

Not every accident results in a lengthy, billion-dollar lawsuit
Negligent AV manufacturers do get punished
Everybody (the public, the press, judges, lawmakers, regulators etc.) has an understandable way to scrutinize the safety of various AVs, both in general and as it relates to a specific accident scenario

This is going to be a non-trivial framework: It will surely have legal and regulatory components. It will probably include ISO-style “process” standards, such as ISO 26262, the SOTIF follow-on, and the expected “SOTIF-for-AVs” follow-on to that. It may contain a formal component and more.
But the central component (tying all others together) is probably going to be a verification system. The presentation will describe how an AV verification system lets you:

Define a comprehensive, continuously-updated library of parameterized scenarios
Run variations of each scenario many times against the AV-in-question, using a proper mix of execution platforms (such as simulation, test tracks etc.)
Evaluate the aggregate of all these scenarios / runs (and any requested subset of it), to transparently understand what was verified (this is called “coverage”) and what “grade” it got

Such a coverage-driven verification system (enhanced by ML-based techniques) is probably our best bet. It will also be a crucial component for improving safety as quickly as
The presentation describes the main attributes of such a system. A Scenario Description Language will be described, and a tools suite to utilize it and create   a path to safer autonomous vehicles is described and presented.   Specific focus is given to regulatory aspects of such a language, and its potential usage as a certification tool.
 

Yoav Hollander

Founder and CTO, Foretellix

Yoav is a world expert and leader in verification. He Invented the “e” verification language and related VLSI verification methodology (later followed by the UVM methodology). Yoav founded Verisity to deliver "e" and related tools.
In the last few years, Yoav has been researching complex-system verification.
He founded Foretellix, and maintains a blog at https://blog.foretellix.com.
13:45 - 14:05

Turning an Automated System into an Autonomous System using Model-Based Design with MATLAB ***VIEW PRESENTATION***

Asaf Moses, Technical Product Manager, Systematics Ltd.

Turning an Automated System into an Autonomous system using Model-Based Design with MATLAB

An autonomous system is a system that functions independently or in a supervised manner and operates under conditions of uncertainty, in an unknown and unpredictable dynamic environment.

Autonomous systems may accumulate new knowledge or adapt themselves to a changing environment. In order to complete their mission, these systems can acquire information from their surroundings, move independently in their environment (real or virtual), avoid dangerous situations and act for a long period time without human intervention.

MathWorks Model-Based Design (MBD) approach, which is based on MATLAB & Simulink, includes many capabilities that allow us to design, simulate and test autonomous systems in a simple and comfortable workflow.

 

This lecture will review a number of capabilities from the autonomous domain. These capabilities will allow us to transform the designed system into an autonomous system that can make decisions independently. During this session, we will present several possibilities for designing an automated system, define and combine different sensors (Sensor Fusion), create perception capabilities in order for the system to better understand its environment, planning and following optimal trajectories, and finally, how our autonomous system can interface with other environments.

This MBD workflow allows us to save a significant amount of time during the development process, up to the prototype stage and on.

Asaf Moses

Technical Product Manager, Systematics Ltd.

Asaf Moses serves as a technical product manager at Systematics Ltd. and responsible for MathWorks Autonomous Systems products family.
As part of this position, Asaf integrates his extensive experience with different companies, in order to accompany and assist their developing process.
Disciplines of expertise: Aeronautics, Control, Physical Modeling and Robotics.
 
Asaf Moses holds a B.Sc. in Mechanical Engineering from Ben-Gurion University and an MBA from the Hebrew University of Jerusalem.

14:05 - 14:25

Driving the Future of Connected Vehicles ***VIEW PRESENTATION***

Yaniv Sulkes, VP Business Development and Marketing, North America & Europe, Autotalks

Driving the future of connected vehicles

***VIEW THE PRESENTATION HERE ***

This presentation will discuss what will happen on our roads until we reach the stage (if at all) when all vehicles will be autonomous. It will present how and why autonomous and manned vehicles must find a way to share our roadways.
Questions will be raised regarding whether or not autonomous vehicles can predict human behaviour and vice-versa, can human drivers anticipate the intentions of autonomous vehicles.
 
Other issues will be raised regarding how to protect vulnerable road users and whether “extra” protection will be needed for motorcycles, bicycles and pedestrians.
V2X will be presented as a solution and additional benefits of this technology will be highlighted including improved safety, mobility, and emissions.
A conclusion will be presented discussing all the main reasons why V2X is an essential technology on the way to full vehicle autonomy.

 

Yaniv Sulkes

VP Business Development and Marketing, North America & Europe, Autotalks

Yaniv brings to Autotalks over 15 years of experience in developing, productizing, marketing and selling innovative technology products. Prior to Autotalks, Yaniv spent a decade serving in various positions at Allot Communications, mainly in product management and marketing. In his last role as the Associate VP of Marketing, he was responsible for Allot’s marketing strategy and execution. Prior to Allot, Yaniv managed a large-scale project at the Electronic Research Department, a top research and development unit of the Israeli Defense Forces. Yaniv holds a B.Sc. in Industrial Engineering & Management and an M.Sc. in Electrical Engineering from Tel-Aviv University.
14:25 - 14:45

How Deep Learning is Enabling Unprecedented Driver and In-Cabin Monitoring using Cameras

Ophir Herbst, CEO, Jungo Connectivity Ltd.

How deep learning is enabling unprecedented driver and in-cabin monitoring using cameras

The topic of driver monitoring is 20 years old by now. With the advancement of autonomous vehicles, interior monitoring is gaining importance:
- For L2 driver monitoring to increase safety by identifying drowsiness, distraction, and personalization of the cabin
- For L3 driver monitoring to enable the handoff control between the semi-autonomous vehicle and the driver
- For L4+ in-cabin monitoring is critical as there is no driver to watch around, e.g to know how many people came in, if they have their seatbelts or not, any medical or distress conditions and so forth
Recent breakthroughs in deep learnign and AI enable us now to do automotive grade driver and cabin monitoring, from various cameras and positions, and running on existing compute inside the vehicle, giving OEMs the critical view of the driver and cabin they need.
In the presentation we will show some of Jungo`s breakthroughs in AIs, and the use cases we enable for upcoming upcoming cars, already licensed by multiple OEMs and Tier 1`s.

Ophir Herbst

CEO, Jungo Connectivity Ltd.

- Serial entrepreneur, with over 20 years of operational and technology-driven experience 

- CEO and founder of Jungo Connectivity, spinoff from Cisco, developing ground-breaking computer-vision automotive driver monitoring product 

- GM at Jungo (acquired by NDS $107m); Executive at NDS (acquired by Cisco, $5b)

- Founder and CEO of Mathtools, specializing in MATLAB compilers and related technologies (acquired by MathWorks)

- B.Sc in Electrical Engineering (cumma sum laude) from the Technion, Israel Institute of Technology

- World-class Bridge player, holding international wins, often representing Israel in international competitions

15:35 - 16:35

Panel- Autonomous Tech – from the Investors Viewpoint

Anat Lea Bonshtien, Panel Moderator, Chairman & Director, Fuel Choices & Smart Mobility Initiative, Prime Minister's Office

Rutie Adar, Head of Samsung Strategy and Innovation Center

Michal Varkat Wolkin, Head of Israel Office, Investments and Innovation, Lear Corporation

Yahal Zilka, Managing Partner, Magma

Danielle Holtz, Director of Business Development, Maniv

Panel Moderator: Anat Lea Bonshtien

Anat Lea Bonshtien

Panel Moderator, Chairman & Director, Fuel Choices & Smart Mobility Initiative, Prime Minister's Office

Dr. Anat Lea Bonshtien serves as the Chairman and Director of the Fuel Choices and Smart Mobility Initiative as of January 2017. Prior to this, she served as the Technology and Regulation Manager of the Initiative. In addition, Dr. Bonshtien serves in the Board of Directors of the Capsula Israeli Smart Transportation Center. She is also a member of the National Science Committee for Energy under the Ministry of Science and Technology. Dr. Bonshtien is a former Mimshak Fellow, a science and policy fellowship program which integrates scientific knowledge with decision making. Dr. Bonshtien holds a MSc and a PhD in Biochemistry from Tel-Aviv University, Israel.

Rutie Adar

Head of Samsung Strategy and Innovation Center

Rutie Adar has been heading Samsung’s Strategy and Innovation Center, Israel, after she joined Samsung in 2010. Within this center Samsung Electronics collaborates and invests in startups focused on core technologies to fuel innovative platforms in growth domains including smart machines and automotive, digital health, internet of things, and artificial intelligence. 
Rutie has been working in the Semiconductors industry during the past 25 years, at both startup and multinational companies. She started her career developing image processing and parallel processing algorithms. In the past 20 years Rutie held various managerial positions in marketing , strategy, planning and business development. She has BSc and MSc in mathematics and computer science from the Hebrew University Jerusalem and MBA from Tel Aviv University. 

Michal Varkat Wolkin

Head of Israel Office, Investments and Innovation, Lear Corporation

Dr. Michal Vakrat Wolkin joined 3M in 2014 as the Global head of Innovation in Israel, as part of 3M’s efforts to expand its investments and technical collaborations internationally. Her role was focused on connecting 3M to the Israeli innovation eco system, with emphasize on Nanotechnology, IOT, Energy, Digital Healthcare and Smart Transportation. She is responsible for R&D collaborations & investments relationships with Israeli startups, academia, government, multinationals, VCs and other industrial partners. Michal joined 3M after living in the Silicon Valley for over 17 years. In her last position Dr. Vakrat Wolkin was the technical Chair of the NASA/ Clean Tech Open Night Rover Challenge since 2012. From 2008, Michal was Better Place director of Battery technologies in Palo Alto where she managed multinational business development and technical teams, working with leading automotive companies, battery manufactures, and tier 1 suppliers. Before joining Better Place, Michal was a researcher. As a member of the research staff at the Palo Alto Research Center (Xerox PARC) Materials and hardware labs, she led a group of researchers developing highly innovative materials in Nanotechnology, Fuel cells, Flexible Electronics and Sensors for drug discovery applications. Prior to that Dr. Vakrat Wolkin was a Post Doc at Xerox PARC. Dr. Wolkin received a PhD in Materials Science, focusing on Nanotechnology and Photonics from the University of Rochester, NY, and an undergraduate degree in Chemical Engineering from The Technicon. Her Post Doc is from Xerox PARC. She holds 10 patent applications, and her research is widely cited (over 2000 citations to her PRL paper on silicon quantum dots). Dr. Vakrat Wolkin currently serves on various industrial review panels related to nanotechnology, batteries, EVs and other technical areas, and is a frequent speaker at industry events. She has been an advisor and reviewer to the USA ARPA-E and other government scientific agencies.

Yahal Zilka

Managing Partner, Magma

Yahal is the co-founder of Magma Venture Partners. Magma is positioned as the leading Venture Capital Fund in Israel, with numerous success stories that Yahal has led. 
Prior to Magma, Yahal served as the CFO of VocalTec Communications from 1995 to 1999, where he was part of the team that pioneered VoIP. Yahal led the company from seed to its public offering on Nasdaq (NASDAQ:CALL).
Yahal is very active with the business development and financing-related activities of the companies that he works with. Yahal led and was on the boards of Waze (acquired by Google), Onavo (acquired by Facebook), Argus (acquired by Continental), DesignArt Networks (acquired by Qualcomm), Phonetic Systems (acquired by Nuance) and Adience (acquired by Teddy Sagi Group). Yahal currently serves as a director on the boards of Valens, Applitools, Iguaz.io, Hola, Scylla DB, Gloat and Magisto, among others. He is also an investor in leading companies such as Vayyar, Fundbox, MeMed and Replay (acquired by Intel).
For the past several years, Yahal has been very involved in the automotive and mobility ecosystem, building strong relationships and focusing on the automotive industry as a major investment opportunity. As a part of Magma, Yahal has invested in seven companies focused on the automotive space and autonomous driving enabling technologies, including Waze, Autotalks, Valens, Innoviz, Argus, Iguazio and JpU. 

Danielle Holtz

Director of Business Development, Maniv

Danielle manages Maniv’s Business Development efforts, overseeing Maniv’s portfolio support and managing the fund’s relationship with its Limited Partners.

Before joining Maniv, Danielle was the Manager of Partnerships at OurCrowd, one of the world’s leading equity crowdfunding platforms. As one of OurCrowd's early employees, Danielle played a critical role in shaping and implementing the company's business development efforts seeing the company through three financing rounds that raised over $200M.  Danielle holds a B.A. in Business Administration & Italian Literature from The Hebrew University of Jerusalem. She also attended the Universita` Di Perugia in Italy where she received a Certificate of Proficiency in Italian.

16:40 - 17:00

Keynote Lecture

Raj Rajkumar, George Westinghouse Professor of Electrical and Computer Engineering; Director, T-SET University Transportation Center; Director, Real-Time and Multimedia Systems Lab

Raj Rajkumar

George Westinghouse Professor of Electrical and Computer Engineering; Director, T-SET University Transportation Center; Director, Real-Time and Multimedia Systems Lab

George Westinghouse Professor of Electrical and Computer Engineering; Director, T-SET University Transportation Center; Director, Real-Time and Multimedia Systems Lab
Rajkumar's interests lie in all aspects of embedded real-time systems and wireless/sensor networks.
His research interests lie in all aspects of embedded real-time systems and wireless/sensor networks. In the domain of embedded real-time systems, my interests include but are not limited to operating systems, scheduling theory, resource management, wired/wireless networking protocols, quality of service management, hardware/software architecture, model-based design tools and power management. In the context of wireless/sensor networks, my research interests span hardware, devices, power-efficient networking protocols, run-time environments, large-scale system architectures, visualization and administrative tools.