On 13 and 14 November, the VDE became the capital for future technologies, on its 125th anniversary!
At the VDE Tec Summit 2018, almost 190 speakers presented and discussed visionary digital ideas. Among them our CEO and President of KI-Verband Jörg Bienert who showed the potential of AI, expectations and how our services can contribute to the market needs.
Also, he emphasized the role of KI Verband and its essential role in supporting Germany in catching up with new technologies.
The highlight: the anniversary evening with keynote speeches on technology policy by Federal Economics Minister Peter Altmaier and Siemens CEO Joe Kaeser and the big birthday party - including a trip into space to the international space station ISS.
Thank you Tecsummit for this exciting journey!
Immediately after the federal government had presented its AI strategy last week, our CEO and President of the German AI Association was giving an interview at the daily evening news on Tagesschau.de. During this interview, also the products of RAYPACK.AI GmbH were shown, emphasizing a range object and person recognition.
The AI strategy is a very comprehensive document with 77 individual measures. However, since concrete implementation plans including associated budget are mainly lacking, a short-term concretization is necessary. "The strategy paper is a good starting point for an intensive process, which must now be started immediately with all participants. Otherwise, this is just a nice collection of ideas."
Have you ever wanted to make objects disappear from an image?
Now, DeepAngel can make this happen. Developed by researches at MIT, DeepAngel is an Artificial Intelligence that erases objects from photographs by using neural network. This algorithm uses Detectron from Facebook AI Research for object recognition and DeepFill for impainting using a neural network. This means that objects to be removed are detected by Detectron and painted over by DeepFill in such a way that they can hardly be distinguished from the background.
If you have any questions about image processing using AI or would like to know how you can add value to your images or videos, please contact us at email@example.com.
We are delighted to invite you to KINNOVATE organized by Euroforum and taking place on 11 and 12 December 2018, in Munich. This event is focusing on the implementation of Artificial Intelligence solutions in business practice. You will get an overview of current applications and find partners for pilot projects and concrete business models.
We are looking forward to presenting our innovative use cases and highly-scalable AI software solutions such as Visual Computing, which among others, is our primary expertise.
We want to contribute to your participation. Therefore we have a special prize for you! Subscribe now to our Newsletter and get the exclusive opportunity to visit the event at a reduced price which amounts up to -600€. Click here to subscribe and stay tuned for our next Newsletter!
Further information about the event can be found here.
Today and in the next two days, the Medientage will take place in Munich. We are pleased to present our services for AI-based automatic indexing of videos and archive material.
You will find us at the booth of Deutsche Telekom. We look forward to your visit!
It was a great honor for aiso-lab that our CSO Dr. Alexander Lorz gave a plenary talk at ELIV (Electronics in Vehicles) about Artificial Intelligence. Other speakers on that conference day (16.10.2018) in Baden-Baden included Prof. Shai Shalev-Shwartz, CTO Mobileye; Dr. Stefan Poledna, member of the executive board TTTech Computertechnik and Gero Schulze Isfort, CSMO Krone Commercial Vehicle.
Many thanks to VDI Wissensforum.
Source: VDI Wissensforum
Over 2,66 million people watched the interview that our CTO Gary Hilgemann gave to the ARD, Germany’s largest TV broadcaster, yesterday. During the interview, the amazing possibilities that AI has created were emphasized as well as the distinct advantages of our services and the tremendous need for investments in development and research in Germany.
RAYPACK.AI and it’s sister company Rebotnix-for the edge hardware-are specializing in Visual Computing and provides customized highly-scalable AI software solutions that power manufacturing, quality control and other fundamental business processes on a multinational level. To achieve that, we are using Rebotnix Gustav-a highly reliable hardware device- as well as other devices, which among others, makes it easy for everyone to use our Visual Analyzer platform.
Our special thanks go to the report team Munich, Fabian Mader and colleagues for their good overview of the status of AI in Germany.
Ever wanted to dance like Michael Jackson or a prima ballerina, but you have two left feet? Thanks to a new AI algorithm, you can now get closer to this dream – at least on video!
This video shows the AI algorithm developed by researchers at the University of California, Berkeley in action. Their approach is described in the paper "Everybody Dance Now".
In this source video, the posture is detected and presented as a so-called "pose stick figure". This posture is then transferred to the person in the target video.
For this process to be temporally coherent, a unique Generative Adversarial Network (GAN) is used. Also, another GAN is used to make a face as detailed and realistic as possible.
This AI algorithm opens up new possibilities for video editing. In the future, anyone can use AI algorithms to perform movements and actions on video that they are not able to do. This advancement in AI makes it almost impossible to distinguish between real and animated video. Except for the many advantages, this new technology also presents some threats. So, for example, you cannot know any more if a person on tape was doing the movement him-/herself.
If you have any questions about image processing using AI, or if you would like to know how to generate more value with your images or videos, feel free to contact us at firstname.lastname@example.org.
The tech giant complements its existing offer with Edge TPU and offers world's first fully-integrated ecosystem to create AI applications
Two years ago, Google introduced its Tensor Processing Units (TPUs) – specialised chips for AI tasks in their data centres. Now, Google offers its cloud expertise as a new Edge TPU. The small AI chip can perform complex Machine Learning (ML) tasks on IoT devices.
The Edge TPU was developed for "inference", the part of machine learning where an algorithm performs the task it was trained to do, such as detecting a defect on a product in a production line. In contrast to that, Google's server-based TPUs are optimized for training machine-learning algorithms.
These new chips are designed to help companies automate tasks such as quality control in factories. For this type of application, an edge device has some advantages over using hardware that has to send data for analysis – requiring a high-speed and stable internet connection. Inference on an edge device is generally safer, more reliable, and delivers faster results. That’s the sales letter, at least.
Of course, Google is not the only company that develops chips for this type of edge application. However, unlike its competitors, Google offers the entire AI stack. Customers can save their data in the Google Cloud, train AI algorithms on TPUs; and then run the trained AI algorithms on the new Edge TPUs. And most likely, customers will also create their machine learning software with TensorFlow - the coding framework developed and operated by Google.
This kind of vertical integration provides tremendous benefits for Google and its customers. On the one hand, Google ensures a perfect connection between the different systems, on the other hand, the customer can program his AI in the most efficient way – using only one platform.
Google Cloud’s Vice President for IoT, Injong Rhee, said: „Edge TPUs are an addition to our cloud TPUs to accelerate the training of ML algorithms in the cloud and then run inference on edge. Sensors are becoming more than just pure data collectors – they can locally take smart decisions in real-time.
Interestingly enough, Google is offering these Edge TPUs as a development kit, which enables the customer to test the devices in their environment. This contradicts Google’s current policy to keep their AI hardware a secret. However, if Google wants to convince customers all over the world to apply only their integrated AI tools, they first need to allow them to test their effectiveness. Hence, offering this development kit is not just a strategy to win new customers, but also shows Google’s efforts to become the leading partner for the company’s AI development efforts.
As soon as this announced development kit is available, our experts will run some tests and provide you with feedback. For further questions, please do not hesitate to contact us at email@example.com.
Source: Google Blogs.
Your child’s first steps, your daughters dancing performance or your son’s fantastic skateboard tricks. These are the moments you will never forget and still, the pass in the blink of an eye. Do you ever wish you would have captured these moments not only on video but in slow motion to enjoy every moment even more?
Nowadays, videos with 240fps (frame-per-second) can be recorded from a mobile phone, while videos with higher frame rates still require a high-speed camera. Also, these high frame rates need an enormous storage capacity and are not feasible in a small mobile device due to their high energy consumption. So how can you ensure to capture and relive these incredible moments often in slow motion – unpredictable and unplanned – without carrying around a sizeable high-speed camera all the time?
Scientists at the University of Massachusetts at Amherst, NVIDIA and the University of California have just found the perfect solution. They successfully conducted a research project in which they used Artificial Intelligence to convert videos with standard frame rates to super slow motion videos. To achieve this, they propose an end-to-end CNN (Convolutional Neural Network) model for frame interpolation that has an explicit sub-network for motion estimation. At first, they used a flow computation CNN to estimate the bidirectional optical flow between the two input frames. In a next step, then two flow fields are linearly fused to refine the approximated flow fields and predict soft visibility maps for interpolation.
Their scientific research has been recently published with the title "Super SloMo: High-Quality Estimation of Multiple Intermediate Frames for Video Interpolation" where they introduce all the necessary steps and experimental results.
Last but not least, speaking about videos, to gain a broader idea about this exceptional achievement, we would recommend you to watch the highly informative video that was recently presented at the renowned CVPR 2018: Computer Vision and Pattern Recognition.