<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:dc="http://purl.org/dc/elements/1.1/">
  <channel>
    <title>Forem: SiaSearch</title>
    <description>The latest articles on Forem by SiaSearch (@siasearch).</description>
    <link>https://forem.com/siasearch</link>
    <image>
      <url>https://media2.dev.to/dynamic/image/width=90,height=90,fit=cover,gravity=auto,format=auto/https:%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Fuser%2Fprofile_image%2F608743%2F09f29483-e4ac-4836-be17-ba54e7dca8d6.jpeg</url>
      <title>Forem: SiaSearch</title>
      <link>https://forem.com/siasearch</link>
    </image>
    <atom:link rel="self" type="application/rss+xml" href="https://forem.com/feed/siasearch"/>
    <language>en</language>
    <item>
      <title>Computer Vision Startups Enhancing Security &amp; Surveillance</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Mon, 13 Sep 2021 07:55:47 +0000</pubDate>
      <link>https://forem.com/siasearch/computer-vision-startups-enhancing-security-surveillance-m7h</link>
      <guid>https://forem.com/siasearch/computer-vision-startups-enhancing-security-surveillance-m7h</guid>
      <description>&lt;p&gt;From common CCTV cameras to autonomous security drones, visual monitoring devices are everywhere. These security systems continuously produce high volumes of footage, much of which sits unused once it's been captured. It's nearly impossible for humans to monitor multiple live security feeds and take proactive action.&lt;/p&gt;

&lt;p&gt;This is where AI comes in - computer vision technology leverages the abundance of visual data to identify what data is useful, what can be ignored, and what demands immediate attention.&lt;/p&gt;

&lt;p&gt;In this article, we at &lt;a href="https://www.siasearch.io/" rel="noopener noreferrer"&gt;SiaSearch&lt;/a&gt; have put together the most promising AI applications for security, as well as the most innovative computer vision startups within each.&lt;/p&gt;

&lt;h2&gt;
  
  
  Video Surveillance
&lt;/h2&gt;

&lt;p&gt;Unlike human personnel, computer vision-based security systems are able to watch security footage tirelessly, monitor everyone in view, identify patterns and any suspicious activity. In recent years, a large number of startups have stepped up to provide AI-powered video surveillance.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://umbocv.ai/" rel="noopener noreferrer"&gt;Umbo&lt;/a&gt; is a Taiwan and San Francisco-based startup that provides cloud-based video security systems for businesses. Umbo's smart security cameras, in combination with proprietary computer vision-based software, autonomously detect and identify suspicious events, such as intrusion, tailgating, and wall-scaling.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.deepsentinel.com/" rel="noopener noreferrer"&gt;Deep Sentinel&lt;/a&gt; has built a similar solution for home security. Whenever a threat is detected, Deep Sentinel streams live video to real human security personnel to remotely intervene via microphone.&lt;/p&gt;

&lt;p&gt;In response to the uptick in mass shootings events, &lt;a href="https://actuate.ai/" rel="noopener noreferrer"&gt;Actuate&lt;/a&gt; (formerly known as Aegis AI) integrates with existing camera feeds to automatically identify anyone who's brandishing a firearm. Once the model identifies a weapon, it alerts security teams and law enforcement.&lt;/p&gt;

&lt;p&gt;"We can detect a weapon before a trigger is pulled," said Ben Ziomek, Actuate co-founder and CTO. "In some instances we can enable a security response before any bullets are fired."&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F34y1cvyy6978ug3vesfv.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F34y1cvyy6978ug3vesfv.png" alt="Actuate"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Access Control
&lt;/h2&gt;

&lt;p&gt;A number of computer vision startups offer innovative solutions to restrict or allow access to certain areas or facilities.&lt;/p&gt;

&lt;p&gt;San Francisco-based &lt;a href="https://www.swiftlane.com/" rel="noopener noreferrer"&gt;Swiftlane&lt;/a&gt;, for instance, uses facial recognition to allow or deny access to offices, apartment complexes, and other physical spaces. The solution employs deep learning and computer vision techniques to provide a single-sign-on using a mobile phone or video intercom. After signing up, users only need to look at the face reading terminal to unlock the doors to areas they are authorized to enter.&lt;/p&gt;

&lt;p&gt;Similarly, &lt;a href="https://www.paravision.ai/" rel="noopener noreferrer"&gt;Paravision's&lt;/a&gt; platform is designed to be used by global security device manufacturers, solution providers, systems integrators, and financial services firms in situations where an error could have profound negative consequences.&lt;/p&gt;

&lt;h2&gt;
  
  
  Checkpoint Security
&lt;/h2&gt;

&lt;p&gt;Computer vision technology has also made great strides in detecting and identifying threats at security screening checkpoints.&lt;/p&gt;

&lt;p&gt;For example, Silicon Valley's &lt;a href="https://www.synapsetechnology.com/" rel="noopener noreferrer"&gt;Synapse Technology&lt;/a&gt; automates security screening, enabling checkpoints worldwide to catch more threats while reducing operating costs and increasing throughput. Their platform, Syntech ONE®, integrates with new and existing checkpoint machines at airports, courthouses, federal buildings, and more.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://www.evolvtechnology.com/" rel="noopener noreferrer"&gt;Evolv Technology's&lt;/a&gt; Edge system uses a combination of camera, facial recognition and millimetre-wave technologies to scan people walking through portable security gates at airports. Machine learning models automatically check for threats, including explosives and firearms, while ignoring non-dangerous items.&lt;br&gt;
Inspecting vehicles can be a challenge for checkpoint security teams. &lt;/p&gt;

&lt;p&gt;Difficult to access, the undercarriage is an ideal spot for adversaries to hide illicit materials such as explosives, weapons, and drugs. &lt;a href="https://www.uveye.com/" rel="noopener noreferrer"&gt;UVeye&lt;/a&gt; is an Israeli startup that provides an automated under vehicle inspection scanner that captures high-resolution images as it scans passing vehicles. Using advanced deep learning algorithms, the scanner is able to detect and flag anomalies in seconds.&lt;/p&gt;

&lt;h2&gt;
  
  
  Theft Detection
&lt;/h2&gt;

&lt;p&gt;Computer vision can also help retailers react to theft and threats as they happen. With roots in MIT's artificial intelligence labs, &lt;a href="https://www.stoplift.com/" rel="noopener noreferrer"&gt;StopLift&lt;/a&gt; analyzes security video and POS data to distinguish between legitimate and fraudulent behavior at checkout. By applying advanced computer vision algorithms to existing camera feeds, StopLift's ScanItAll system tracks items that pass through the checkout lane, associate them with POS, and flag any suspicious activity.&lt;/p&gt;

&lt;p&gt;Developed by Japanese telecommunications company NTT East and startup Earth Eyes, AI Guardman is an automated security camera designed to catch shoplifters in the act. Based on open-source technology developed at Carnegie Mellon University, AI Guardman scans live video streams from cameras in convenience stores and supermarkets, tracking every customer inside. When a threat is detected, the system sends an alert to shop staff in real time.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4rsp0hhvu20g2sqnu2tn.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2F4rsp0hhvu20g2sqnu2tn.png" alt="AI Guardman"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Public Health &amp;amp; Safety
&lt;/h2&gt;

&lt;p&gt;Advances in computer vision technology can also be used to address common public safety concerns, from cutting crime rates to slowing down the spread of infectious diseases.&lt;/p&gt;

&lt;p&gt;Atlanta-based startup &lt;a href="https://www.flocksafety.com/" rel="noopener noreferrer"&gt;Flock Safety&lt;/a&gt; aims to reduce crime in America by 25% by using computer vision to improve public safety in neighborhoods. Their automated license plate reader (ALPR) software, FlockOS, combines character recognition with computer vision and machine learning to provide real-time insights to crime prevention authorities in over 1200 US cities. It uses patented Vehicle Fingerprint™ technology that identifies a vehicle even if it's modified or if its license plate is missing or covered.&lt;/p&gt;

&lt;p&gt;Furthermore, in the wake of the COVID-19 pandemic, many public facilities have equipped existing security cameras with AI-based software that can track compliance with health guidelines.&lt;/p&gt;

&lt;p&gt;Reducing the risk of infection in stores is a major priority for brick-and-mortar retailers in particular. Retail analytics platform &lt;a href="https://auravision.ai/covid-solutions/" rel="noopener noreferrer"&gt;Aura Vision&lt;/a&gt; provides a suite of COVID-focused solutions to promote in-store safety, from features that monitor face mask compliance to heat maps to visualize high traffic areas since the last clean.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftobb2i8qkaatfhwiajpf.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Ftobb2i8qkaatfhwiajpf.png" alt="Aura Vision"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Homeland Security
&lt;/h2&gt;

&lt;p&gt;Not only has computer vision proven its practical value for physical security on private property and public spaces, it has also demonstrated value on a national scale, with applications ranging from environmental monitoring to use in military systems.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://shield.ai/" rel="noopener noreferrer"&gt;Shield AI&lt;/a&gt; is a company that works with federal, state, and local departments and agencies to deliver next generation surveillance systems. Their first product, Nova, is a Hivemind-powered drone that searches buildings while simultaneously streaming video and generating maps.&lt;/p&gt;

&lt;p&gt;Finally, &lt;a href="https://orbitalinsight.com/" rel="noopener noreferrer"&gt;Orbital Insight&lt;/a&gt; specializes in applying computer vision to geo-analytics. The company uses satellites, drones, balloons, UAV footage, and geolocation data from mobile phones to analyze human activities and provide businesses and governments with key behavioral insights and address security concerns.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuw65ywm6sq3eplatd7v6.png" class="article-body-image-wrapper"&gt;&lt;img src="https://media.dev.to/dynamic/image/width=800%2Cheight=%2Cfit=scale-down%2Cgravity=auto%2Cformat=auto/https%3A%2F%2Fdev-to-uploads.s3.amazonaws.com%2Fuploads%2Farticles%2Fuw65ywm6sq3eplatd7v6.png" alt="Infographic: Computer Vision Startups in Security"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Visual Data Management for the Security Industry
&lt;/h2&gt;

&lt;p&gt;In the security sector, the challenge is not data acquisition, but effective data management. The majority of security camera recordings aren't useful for the relevant business or computer vision functions, causing extremely high redundancy and low information density.&lt;br&gt;
SiaSearch makes it easy to select subsets of highly quality training data, helping you build better ML models at lower costs.&lt;/p&gt;

&lt;p&gt;With SiaSearch's lightweight API, users can:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Automatically index, structure, and evaluate raw data captured by security cameras and sensors&lt;/li&gt;
&lt;li&gt;Visualize data and analyze model performance&lt;/li&gt;
&lt;li&gt;Search and access all security footage across all events and attributes&lt;/li&gt;
&lt;li&gt;Identify rare edge cases and curate training datasets&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Interested in learning more? &lt;a href="https://www.siasearch.io/request-a-demo" rel="noopener noreferrer"&gt;Reach out to SiaSearch&lt;/a&gt; for a free proof of concept.&lt;/p&gt;

</description>
      <category>security</category>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>computerscience</category>
    </item>
    <item>
      <title>10 Computer Vision Startups Disrupting the Retail Industry</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Mon, 09 Aug 2021 01:02:01 +0000</pubDate>
      <link>https://forem.com/siasearch/10-computer-vision-startups-disrupting-the-retail-industry-2lck</link>
      <guid>https://forem.com/siasearch/10-computer-vision-startups-disrupting-the-retail-industry-2lck</guid>
      <description>&lt;p&gt;&lt;em&gt;This article was originally published at: &lt;a href="https://www.siasearch.io/blog/computer-vision-startups-retail-industry"&gt;https://www.siasearch.io/blog/computer-vision-startups-retail-industry&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;Online retail has been growing steadily for years with no sign of stopping. Especially amid movement restrictions induced by the COVID-19 pandemic, &lt;a href="https://news.un.org/en/story/2021/05/1091182"&gt;research shows that global online sales jumped to $26.7 trillion&lt;/a&gt; in 2020. With the rise of ecommerce, one thing is abundantly clear: brick-and-mortar retailers need to innovate if they want to stay competitive. &lt;/p&gt;

&lt;p&gt;The use of AI technologies like computer vision is rapidly increasing in the retail industry. &lt;a href="https://www.siasearch.io/retail"&gt;AI-enhanced retail&lt;/a&gt; holds the promise to eliminate operational inefficiencies and provide shoppers with frictionless in-store experiences. In this article, we’ve put together a list of the most innovative computer vision startups in the retail space.&lt;/p&gt;

&lt;h2&gt;
  
  
  The Future of AI in Retail
&lt;/h2&gt;

&lt;p&gt;In recent years, an increasing number of retail companies have started to quietly transform physical stores. Walmart, for instance, began &lt;a href="https://corporate.walmart.com/newsroom/2019/04/25/walmarts-new-intelligent-retail-lab-shows-a-glimpse-into-the-future-of-retail-irl"&gt;installing an array of sensors, cameras and processors&lt;/a&gt; to monitor inventory levels, perform automated product quality checks, and more.&lt;/p&gt;

&lt;p&gt;Many industry giants have followed suit by employing similar approaches to drive in-store efficiency, better logistics, prevent theft, and more. Research shows that the retail AI market is growing fast—according to a recent RIS News report, only 3% of retailers were utilizing computer vision technology at the end of 2020, yet an additional 40% had plans to deploy new solutions within the next year. &lt;/p&gt;

&lt;p&gt;Computer vision solutions present retailers with ample opportunity to boost operations as well as enhance the shopping experience for customers. Among the most promising applications of computer vision include inventory management, loss prevention, automated checkout, and behavioral analytics. From employee-free shops to in-store surveillance, here are several computer vision startups disrupting the retail industry.&lt;/p&gt;

&lt;h2&gt;
  
  
  Computer vision startups in retail
&lt;/h2&gt;

&lt;p&gt;&lt;strong&gt;Inventory management&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://goradar.com/"&gt;RADAR&lt;/a&gt;: RADAR is a fully integrated hardware and software solution to automate inventory management using RFID and computer vision techniques. Their mission is to streamline inventory management via automated inventory counts, improved in-store replenishment and instantaneous customer stock checks.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://traxretail.com/resources/trax-computer-vision-platform/"&gt;Trax&lt;/a&gt;: Singapore-based startup Trax provides an in-store solution that uses a combination of computer vision models and hardware to keep track of their inventory in real time. This solution ensures out-of-stock items are repurchased efficiently, while expired items are pulled off from the shelves. The company holds 23 patents on its technology and can analyze images from phones, in-store cameras, and grocery store robots.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Automated checkout&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://standard.ai/"&gt;Standard.ai&lt;/a&gt;: Previously known as Standard Cognition, Standard.ai’s automated checkout solution is made to fit with retailers’ existing stores and technology. They boast an easy to install camera-first solution that doesn’t employ the use of turnstiles or gates. Standard doesn't use any facial recognition or biometrics, and all deployments are on-premise to ensure maximum performance and security for retailers and shoppers alike.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://trigo.tech/"&gt;Trigo&lt;/a&gt;: Using proprietary algorithms and affordable off-the-shelf sensor kits, Tel Aviv-based Trigo allows retailers to analyze anonymized shoppers’ movements and product choices in real time. The system automatically compiles selected items into a virtual shopping list, enabling shoppers to leave without going through a traditional checkout line.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://www.accelrobotics.com/"&gt;Accel Robotics&lt;/a&gt;: Accel Robotics provides checkout-free shopping experiences across existing and new store formats with its patented camera-based AI system. They recently launched Valet Market, a completely automated convenience storefront without cashiers or checkout kiosks.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Loss prevention&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="https://www.stoplift.com/"&gt;StopLift&lt;/a&gt;: With roots in MIT’s artificial intelligence labs, StopLift analyzes security video and POS data to distinguish between legitimate and fraudulent behavior at checkout. By applying advanced computer vision algorithms to existing camera feeds, StopLift’s ScanItAll system is capable of tracking items that pass through the checkout lane, associate them with POS, and flag suspicious activity as it happens.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://vaak.co/vaakeye-store/"&gt;Vaak&lt;/a&gt;: Japanese startup Vaak provides a cloud-based computer vision system that monitors retail security camera footage for suspicious behavior. Already deployed in over 50 stores within Japan, VaakEye analyzes movement at more than 100 points across the body, automatically weighing behavior for suspiciousness. Once a customer reaches a certain threshold, the system sends an alert, along relevant video clips, to the appropriate staff member.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;strong&gt;Behavioral analytics&lt;/strong&gt;&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;a href="http://deepnorth.com/"&gt;Deep North&lt;/a&gt;: Deep North provides an analytics platform that builds real-time video intelligence for retailers based on video data from CCTV and other cameras that those retailers already use. Deep North’s proprietary technology captures parameters as daily entries and exits, customer occupancy, queue times, conversions and more.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://advertima.com/"&gt;Advertima&lt;/a&gt;: Based on information captured by visual sensors, Advertima’s platform provides retailers with a real-time view of what’s going on in physical stores as shoppers move through the space. The platform claims to only process minimal anonymized data, without storing any recordings or personal information for future use.&lt;/li&gt;
&lt;li&gt;
&lt;a href="https://cosmose.co/"&gt;Cosmose&lt;/a&gt;: Cosmose provides a data analytics platform that analyzes foot traffic in physical stores to help predict customer behavior. They offer 3 main products: Cosmose Analytics, which tracks customers’ movements inside physical stores; Cosmose AI, a data analytics and prediction platform to help retailers create marketing campaigns and drive sales; and Cosmose Media, for targeting online ads.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;Real-time visibility is essential to operating brick-and-mortar retail. That’s why more and more retailers are employing computer vision in an effort to increase operational efficiency, better the customer experience, and gain an edge over competitors. &lt;/p&gt;

&lt;p&gt;Early adopters are already seeing great results—according to estimates from RBC Capital Markets analysts, cashierless Amazon Go stores bring in about 50% more revenue on average than typical convenience stores.&lt;/p&gt;

&lt;h2&gt;
  
  
  Visual data management for retail
&lt;/h2&gt;

&lt;p&gt;Companies that build computer vision solutions for retail are constantly building and growing their ML training data sets. Today, most companies have to rely on internal tools or manual solutions like spreadsheets to do this. &lt;/p&gt;

&lt;p&gt;SiaSearch helps retail companies to simplify and speed up this process with a lightweight API that simplifies data exploration, visualization and selection. As a result, companies can reduce annotation costs and increase model performance.&lt;/p&gt;

&lt;p&gt;Interested in learning more? &lt;a href="https://www.siasearch.io/request-a-demo"&gt;Reach out to the SiaSearch team&lt;/a&gt; for a free proof of concept.&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>startup</category>
      <category>computerscience</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>15 Best Open-Source Autonomous Driving Datasets</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Wed, 09 Jun 2021 07:05:45 +0000</pubDate>
      <link>https://forem.com/siasearch/15-best-open-source-autonomous-driving-datasets-2e81</link>
      <guid>https://forem.com/siasearch/15-best-open-source-autonomous-driving-datasets-2e81</guid>
      <description>&lt;p&gt;In recent years, more and more companies and research institutions have made their &lt;a href="https://www.siasearch.io/open-data"&gt;autonomous driving datasets&lt;/a&gt; open to the public. However, the best datasets are not always easy to find, and scouring the internet for them takes time.&lt;/p&gt;

&lt;p&gt;To help, we at SiaSearch have put together a list of the top 15 open datasets for autonomous driving. The resources below collectively contain millions of data samples, many of which are already annotated. We hope this list provides you with a solid starting point for learning more about the field, or for starting your own autonomous driving project.&lt;/p&gt;

&lt;h2&gt;
  
  
  Top Open Datasets for Autonomous Driving Projects
&lt;/h2&gt;

&lt;ol&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;A2D2 Dataset&lt;/strong&gt;&lt;br&gt;
The Audi Autonomous Driving Dataset (A2D2) features over 41,000 labeled with 38 features. Around 2.3 TB in total, A2D2 is split by annotation type (i.e. semantic segmentation, 3D bounding box).&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;ApolloScape Dataset&lt;/strong&gt;&lt;br&gt;
ApolloScape is an evolving research project that aims to foster innovation across all aspects of autonomous driving, from perception to navigation and control. Via their website, users can explore a variety of simulation tools and over 100K street view frames, 80k lidar point cloud and 1000km trajectories for urban traffic.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Argoverse Dataset&lt;/strong&gt;&lt;br&gt;
The Argoverse dataset includes 3D tracking annotations for 113 scenes and over 324,000 unique vehicle trajectories for motion forecasting.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Berkeley DeepDrive Dataset&lt;/strong&gt;&lt;br&gt;
Also known as BDD 100K, the DeepDrive dataset gives users access to 100,000 annotated videos and 10 tasks to evaluate image recognition algorithms for autonomous driving. The dataset represents more than 1000 hours of driving experience with more than 100 million frames, as well as information on geographic, environmental, and weather diversity.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;CityScapes Dataset&lt;/strong&gt;&lt;br&gt;
CityScapes is a large-scale dataset focused on the semantic understanding of urban street scenes in 50 German cities. It features semantic, instance-wise, and dense pixel annotations for 30 classes grouped into 8 categories. The entire dataset  includes 5,000 annotated images with fine annotations, and an additional 20,000 annotated images with coarse annotations.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Comma2k19 Dataset&lt;/strong&gt;&lt;br&gt;
This dataset includes 33 hours of commute time recorded on highway 280 in California. Each 1-minute scene was captured on a 20km section of highway driving between San Jose and San Francisco. The data was collected using comma EONs, which features a road-facing camera, phone GPS, thermometers and a 9-axis IMU. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Google-Landmarks Dataset&lt;/strong&gt;&lt;br&gt;
Published by Google in 2018, the Landmarks dataset is divided into two sets of images to evaluate recognition and retrieval of human-made and natural landmarks. The original dataset contains over 2 million images depicting 30 thousand unique landmarks from across the world. In 2019, Google published Landmarks-v2, an even larger dataset with 5 million images and 200k landmarks.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;KITTI Vision Benchmark Suite&lt;/strong&gt;&lt;br&gt;
First released in 2012 by Geiger et al, the &lt;a href="https://www.siasearch.io/kitti-dataset/"&gt;KITTI dataset&lt;/a&gt; was released with the intent of advancing autonomous driving research with a novel set of real-world computer vision benchmarks. One of the first ever autonomous driving datasets, KITTI boasts over 4000 academic citations and counting.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Level 5 Open Data&lt;/strong&gt;&lt;br&gt;
Published by popular rideshare app Lyft, the Level5 dataset is another great source for autonomous driving data. It includes over 55,000 human-labeled 3D annotated frames, surface map, and an underlying HD spatial semantic map that is captured by 7 cameras and up to 3 LiDAR sensors that can be used to contextualize the data.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;nuScenes Dataset&lt;/strong&gt;&lt;br&gt;
Developed by Motional, the &lt;a href="https://www.siasearch.io/nuscenes-dataset/"&gt;nuScenes dataset&lt;/a&gt; is one of the largest open-source datasets for autonomous driving. Recorded in Boston and Singapore using a full sensor suite (32-beam LiDAR, 6 360° cameras and radars), the dataset contains over 1.44 million camera images capturing a diverse range of traffic situations, driving maneuvers, and unexpected behaviors.&lt;/p&gt;&lt;/li&gt;
&lt;/ol&gt;

&lt;p&gt;Looking for more datasets? Read the entire blogpost at &lt;a href="https://www.siasearch.io/blog/best-open-source-autonomous-driving-datasets"&gt;https://www.siasearch.io/blog/best-open-source-autonomous-driving-datasets&lt;/a&gt;. &lt;/p&gt;

</description>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>database</category>
      <category>autonomousdriving</category>
    </item>
    <item>
      <title>SiaSearch partners with Virtual Mechanics Corporation (VMC) to accelerate ADAS development in Japan</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Mon, 31 May 2021 02:17:47 +0000</pubDate>
      <link>https://forem.com/siasearch/siasearch-partners-with-virtual-mechanics-corporation-vmc-to-accelerate-adas-development-in-japan-2lbp</link>
      <guid>https://forem.com/siasearch/siasearch-partners-with-virtual-mechanics-corporation-vmc-to-accelerate-adas-development-in-japan-2lbp</guid>
      <description>&lt;p&gt;&lt;a href="https://www.siasearch.io/"&gt;SiaSearch&lt;/a&gt;, a Berlin-based AI startup, has announced their partnership with &lt;a href="https://vmc.jp/"&gt;Virtual Mechanics Corporation (VMC)&lt;/a&gt; to accelerate ADAS development in Japan. Virtual Mechanics Corporation has over 20 years of experience and established relationships within the Japanese automotive industry. As the sole distributor of the SiaSearch product in Japan, Virtual Mechanics Corporation will promote the adoption of SiaSearch in the Japanese market.&lt;/p&gt;

&lt;p&gt;Via web-based GUI or programmatic API, SiaSearch makes it 10x easier and faster for developers to explore, understand, and share large amounts of visual data.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Structure - Automatically index, structure, and evaluate raw sensor data on a petabyte-scale, based on semantic attributes and keywords from a catalog of over 50 events and attributes&lt;/li&gt;
&lt;li&gt;Analyze - Quickly visualize your data, making full use of the extracted attributes and run more targeted analyses to improve your models.&lt;/li&gt;
&lt;li&gt;Search - Use custom attributes to efficiently find and select the data from your entire data lake.&lt;/li&gt;
&lt;li&gt;Collaborate - Save, edit, version, comment and share frames, sequences or objects with colleagues or 3rd parties.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;“At Virtual Mechanics Corporation, we are laser focused on driving innovation within the mobility industry,” said Eiji Takita, CEO of Virtual Mechanics Corporation. “In working with SiaSearch, we have secured a partner with deep technical capabilities to help accelerate autonomous driving innovation in Japan.”&lt;/p&gt;

&lt;p&gt;“Our distribution partnership with Virtual Mechanics Corporation opens up new and exciting opportunities to accelerate autonomous driving innovation in the Japanese market,” said Clemens Viernickel, CEO of SiaSearch. “We are beyond excited to be working with some of the globally leading automotive companies in Japan through this arrangement. The partnership represents an important milestone for us, and we will continue to pave the path of adoption of SiaSearch abroad.”&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;About SiaSearch&lt;/strong&gt;&lt;br&gt;
SiaSearch is a Berlin-based AI company on a mission to power automated mobility solutions of tomorrow with a truly scalable data infrastructure. The category-leading sensor management platform for automated driving, SiaSearch is currently in use by leading automotive OEMs, Tier 1 suppliers, and global technology companies.&lt;br&gt;
For more information, visit &lt;a href="https://www.siasearch.io/"&gt;https://www.siasearch.io/&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;About Virtual Mechanics Corporation (VMC)&lt;/strong&gt;&lt;br&gt;
Virtual Mechanics Corporation is working on vehicle dynamics simulation software to simulate driving in a virtual environment and its application technology. Leading customers, the developer and partner companies, Virtual Mechanics Corporation plays a pivotal role in innovation, allowing customers to expedite their research and development.&lt;br&gt;
For more details, please visit &lt;a href="https://vmc.jp/siasearch/"&gt;https://vmc.jp/siasearch/&lt;/a&gt; (information in Japanese)&lt;/p&gt;

</description>
      <category>datascience</category>
      <category>machinelearning</category>
    </item>
    <item>
      <title>KITTI on SiaSearch - Our first public product unveil for researchers</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Fri, 14 May 2021 02:14:17 +0000</pubDate>
      <link>https://forem.com/siasearch/kitti-on-siasearch-our-first-public-product-unveil-for-researchers-2f3i</link>
      <guid>https://forem.com/siasearch/kitti-on-siasearch-our-first-public-product-unveil-for-researchers-2f3i</guid>
      <description>&lt;p&gt;Today, we are glad to announce the release of a public version of SiaSearch based on the popular &lt;a href="https://www.siasearch.io/kitti-dataset/"&gt;KITTI dataset&lt;/a&gt;. We would like to let researchers use SiaSearch’s power in order to immensely simplify data searchability. We hope that the use of KITTI would allow them to seamlessly extract interesting insights from one of the most popular datasets in AV/ADAS. Deployed on KITTI, we want to make a subset of the features of SiaSearch accessible to researchers all around the world.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--D0WXoMbB--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/f7m9vawoqh2pgptd8c9p.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--D0WXoMbB--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/f7m9vawoqh2pgptd8c9p.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;SiaSearch allows to efficiently search through recorded driving data. It is made available to researchers on the KITTI dataset.&lt;/p&gt;

&lt;h2&gt;
  
  
  About SiaSearch
&lt;/h2&gt;

&lt;p&gt;Before we dive into the demo walkthrough, let’s quickly recap the need and value of &lt;a href="https://www.siasearch.io/"&gt;SiaSearch&lt;/a&gt;. As we explained in detail in our previous post, data volumes in the AV/ADAS domain are exploding at an increasing pace with a single vehicle producing upto 10TBs per hour today. Therefore, the ability to search, analyze and prioritize that data in an efficient manner becomes fundamental to its utilization. SiaSearch allows users to process large quantities of multimodal automotive data and extract queryable metadata. Using this metadata developers can easily find complex situations encountered by the vehicle, ranging from lane changes to overtaking to unsafe braking. With fast search, we reduce the time wasted on repetitive data tasks by instantly connecting engineers with relevant data. Addressing data access further enables us to enable smarter data retention decisions and therefore significantly optimize infrastructure usage.&lt;/p&gt;

&lt;h2&gt;
  
  
  SiaSearch Features
&lt;/h2&gt;

&lt;p&gt;In order to allow you to experience SiaSearch’s abilities, let’s quickly walk through the most important features and functions:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;
&lt;strong&gt;Querying&lt;/strong&gt; — In SiaSearch there are two methods with which you can query for the data you want: The visual (default) and the code interface. The code query works like any API call statement would, whereas the visual query offers a visually rich interface to make the selection of extractors and search extremely intuitive. Users can select an extractor from a variety of categories. The user can select multiple extractors and then apply search to obtain the results. The code and visual query interfaces are directly linked and updated accordingly.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Similarity Search&lt;/strong&gt; — When a user comes across an interesting data segment, they might want to find similar clips of data. Since we have high dimensional multimodal data, this would be extremely hard to do manually. Therefore we integrated an automated similarity search, based on unsupervised clustering of the previously extracted metadata to fetch the most relevant and similar data segments. It is important to note that similarity search is NOT a query. In order to find the same clips, queries can be sent. Similarity analysis is more of an explorative tool, taking into account the metadata clusters.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Replay/Adjust Clip&lt;/strong&gt; — A user might also want to view the data right before or after a certain situation (e.g. a left turn) in order to contextualize the vehicle behavior. This is why we allow the user to change the length of the clip according to their needs. Clip re-adjustment can be achieved by clicking on the adjust clip button in the replay viewer. This opens up the slider so that the user can specify the new start and end points and apply the changes to get an updated clip.&lt;/li&gt;
&lt;li&gt;
&lt;strong&gt;Export&lt;/strong&gt; — To empower the user with all our metadata insights we allow export of selected data as parquet files. Having this metadata can allow the user to do further analysis or to directly use it any application. After collecting several segments of choice the user can go to the export page, select the parts of metadata which should be exported and obtain the parquet file.&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  SiaSearch Workflows
&lt;/h2&gt;

&lt;p&gt;SiaSearch can unlock many different workflows in both academic and commercial settings ranging from scenario extraction for AV stack simulation testing to data sharing/collaborative analysis. However, one of the most important ones is filtering data for annotation. This is shown in the video below to enable contextual understanding of the product.&lt;/p&gt;

&lt;h2&gt;
  
  
  Try it today!
&lt;/h2&gt;

&lt;p&gt;If you are interested in using &lt;a href="https://www.siasearch.io/kitti-dataset/"&gt;KITTI on SiaSearch&lt;/a&gt; for your research, &lt;a href="http://public.sia-search.com/"&gt;click here to get started&lt;/a&gt;. Our team will set up a dedicated account for you so that you can see SiaSearch’s abilities for yourself.&lt;/p&gt;

</description>
      <category>opensource</category>
      <category>machinelearning</category>
      <category>computerscience</category>
      <category>datascience</category>
    </item>
    <item>
      <title>The Best Data Curation Tools for Computer Vision in 2021</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Mon, 03 May 2021 01:12:31 +0000</pubDate>
      <link>https://forem.com/siasearch/the-best-data-curation-tools-for-computer-vision-in-2021-21bl</link>
      <guid>https://forem.com/siasearch/the-best-data-curation-tools-for-computer-vision-in-2021-21bl</guid>
      <description>&lt;p&gt;The term ‘curation’ is commonly associated with museums or libraries, not data science. However, much like the work that’s done on rare paintings or books, data curation tools make the most important data easily accessible to engineers as they build complex machine learning models.&lt;/p&gt;

&lt;p&gt;Without curation, data is difficult to find, analyze, and interpret. Data curation tools provide meaningful insights and enduring access to all your data in one place. In this article, we’ll dive into the importance of data curation for computer vision specifically, as well as review the top data curation tools on the market today.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is data curation?
&lt;/h2&gt;

&lt;p&gt;Data curation is the act of organizing, enhancing, and preserving data for future use. In machine learning, data curation describes the management of data throughout its lifecycle: from its collection and initially storage, to the time it is archived for future re-use.&lt;/p&gt;

&lt;p&gt;This process is all the more important for computer vision engineers, who deal with massive amounts of visual data on a daily basis. Instead of using manual methods such as writing ETL jobs to extract insights, data curation tools provide a streamlined way to access the right data whenever you need to.&lt;/p&gt;

&lt;h2&gt;
  
  
  The importance of data curation for machine learning
&lt;/h2&gt;

&lt;p&gt;Under the hood, data curation tools directly influence computer vision model performance. Using data curation tools, engineers can get a better understanding of the data they’ve collected, identify the most important subsets and edge cases, and curate custom training datasets to feed back into their models.&lt;/p&gt;

&lt;p&gt;The best data curation tools enable you to:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Visualize large scale data&lt;/strong&gt;: Make it easy to obtain insights on key metrics, as well as the general distribution and diversity of your datasets regardless of sensor type and format. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Enable data discovery and retrieval&lt;/strong&gt;: Quickly search, filter, and sort through the entire data lake by making all features queryable and easily accessible. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Curate diverse scenarios&lt;/strong&gt;: Identify the most interesting segments within your dataset, and manipulate them within the tool to create completely customized training sets.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Seamlessly integrate&lt;/strong&gt;: The tool should fit well within your existing workflows and toolset.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  What are the best data curation tools for computer vision?
&lt;/h2&gt;

&lt;p&gt;With an overwhelming amount of AI products and platforms popping up year after year, how do you know which will provide the most value? Based on our experience, we are sharing our honest reviews of the top tools, hoping that this will be of use for engineers searching for a data curation solution.&lt;/p&gt;

&lt;p&gt;Read on below to find out which data curation tool is the best fit for your computer vision project.&lt;/p&gt;

&lt;h3&gt;
  
  
  Aquarium Learning
&lt;/h3&gt;

&lt;p&gt;Aquarium is a data management platform that aims to make it easy to identify labeling errors and model failures. With Aquarium, users can version and combine model predictions with their ground truth. &lt;/p&gt;

&lt;p&gt;Aquarium is especially focused on curating and maintaining training datasets, catering less to raw data management use cases. This is because data exploration in Aquarium is predominantly tied to model predictions and ground truth labels. &lt;/p&gt;

&lt;p&gt;Users can access Aquarium via their cloud platform or API. However, they currently do not offer on-premise or VPC deployments, and there are no external integrations.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Wide range of use cases&lt;/strong&gt; - Aquarium supports image, 3D, audio, and text data. They also support multiple annotation types, such as classification, detection, and segmentation.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Interactive model evaluation&lt;/strong&gt; - Users can manipulate evaluation thresholds and obtain interactive visualizations to obtain required samples quickly.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Collaborative features&lt;/strong&gt; - Users can collaborate with each other on the Aquarium platform to build data subsets, associate them with issues, and identify new data for annotation.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  Scale Nucleus
&lt;/h3&gt;

&lt;p&gt;Launched in late 2020 by Scale, Nucleus is one of the newest data curation tools to hit the market. The Nucleus platform allows users to collaboratively search through image data for model failures. As of now, Nucleus only supports image data, with no support for 3D sensor fusion, video, or text data.&lt;/p&gt;

&lt;p&gt;Users can access Nucleus via their cloud platform, API or Python SDK. Currently, Nucleus does not support on-premise deployability.&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Visual similarity&lt;/strong&gt; - Users can search for visually similar images based on one or multiple base samples and associate custom tags with them.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Metadata schemas&lt;/strong&gt; - Using the Nucleus SDK, users can create flexible metadata schemas. Nucleus provides smart methods to detect and create schemas using the annotation format provided. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Model versioning&lt;/strong&gt; - Users can create model entities and associate corresponding runs with them. Hence, models can be versioned based on runs (dataset &amp;amp; predictions).&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h3&gt;
  
  
  SiaSearch
&lt;/h3&gt;

&lt;p&gt;&lt;a href="https://www.siasearch.io/"&gt;SiaSearch&lt;/a&gt; is a data management platform for computer vision data. Consisting of a scalable metadata catalog and query engine, SiaSearch enables developers to easily search through visual data, add metadata to frames and sequences, as well as assemble custom subsets of data for training or testing. &lt;/p&gt;

&lt;p&gt;With deep roots in autonomous driving, the SiaSearch platform is used by many OEMs, Tier 1s and tech companies. Aside from autonomous driving, SiaSearch also has solutions for robotics, retail, &lt;a href="https://www.siasearch.io/usecases"&gt;and more&lt;/a&gt;. &lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Specialized in sensor data&lt;/strong&gt; - One of the only tools that can support 3D sensor fusion data, SiaSearch can analyze large volumes of unstructured sensor data, providing insights at the frame and sequence level. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Auto-tagging capabilities&lt;/strong&gt; - SiaSearch employs a large catalog of pre-trained extractors to automatically add frame-level, contextual metadata to raw data. Additionally, SiaSearch provides a toolbox for quick extractor development, allowing developers to integrate their own extractors.&lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Fast performance&lt;/strong&gt; - The SiaSearch platform features a unique, proprietary architecture that combines numeric and sequence-based queries to enable noticeably faster performance. &lt;/p&gt;&lt;/li&gt;
&lt;li&gt;&lt;p&gt;&lt;strong&gt;Flexible workflows &amp;amp; integrations&lt;/strong&gt; - Users can access SiaSearch via their web-based GUI or programmatic API. SiaSearch also supports cloud or on-premise deployment for enterprise users.&lt;/p&gt;&lt;/li&gt;
&lt;/ul&gt;

&lt;h2&gt;
  
  
  Interested in data curation?
&lt;/h2&gt;

&lt;p&gt;The right data curation tool can dramatically reduce the time spent on manual processes, allowing engineers to focus on what really matters - building great models.&lt;/p&gt;

&lt;p&gt;If you’d like to hear more about what we’re doing at SiaSearch, reach out to us at &lt;a href="mailto:hi@siasearch.io"&gt;hi@siasearch.io&lt;/a&gt;, or visit &lt;a href="https://www.siasearch.io/"&gt;the SiaSearch website&lt;/a&gt; to learn more.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;Read the full list on our blog: &lt;a href="https://www.siasearch.io/blog/best-data-curation-tools-for-computer-vision"&gt;https://www.siasearch.io/blog/best-data-curation-tools-for-computer-vision&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>computerscience</category>
      <category>database</category>
    </item>
    <item>
      <title>Effortlessly explore the nuScenes dataset with SiaSearch</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Mon, 12 Apr 2021 01:50:22 +0000</pubDate>
      <link>https://forem.com/siasearch/effortlessly-explore-the-nuscenes-dataset-with-siasearch-275</link>
      <guid>https://forem.com/siasearch/effortlessly-explore-the-nuscenes-dataset-with-siasearch-275</guid>
      <description>&lt;h2&gt;
  
  
  A guide to better access, explore, and understand unstructured sensor data for autonomous driving development
&lt;/h2&gt;

&lt;p&gt;To accelerate the speed of autonomous vehicle adoption, an increasing number of organizations and individuals are making their projects available to the public. Open data is fueling commercial and technological advancement in autonomous driving—one of most well known resources being the &lt;a href="https://www.nuscenes.org/"&gt;nuScenes dataset&lt;/a&gt;. &lt;/p&gt;

&lt;p&gt;Developed by the team at Motional (formerly nuTonomy), nuScenes is one of the most popular open-source datasets for autonomous driving. The nuScenes dataset enables researchers to study a wide range of urban driving situations using data captured by the full sensor suite of a self-driving car. The first dataset of its kind, nuScenes was a key player in cultivating a culture of data sharing and collaboration within the mobility industry.&lt;/p&gt;

&lt;p&gt;To further advance this mission, Motional recently partnered with Berlin-based startup &lt;a href="https://www.siasearch.io/"&gt;SiaSearch&lt;/a&gt; to introduce a completely new way to interact with nuScenes—by using a data curation platform to delve deeper into the data than ever before.&lt;/p&gt;

&lt;p&gt;This guide will walk you through how to better access, explore and understand the nuScenes dataset with the SiaSearch platform.&lt;/p&gt;

&lt;h2&gt;
  
  
  What is the nuScenes Dataset?
&lt;/h2&gt;

&lt;p&gt;The nuScenes dataset is one of the largest public datasets for autonomous driving. The dataset contains a rich library of meticulously hand-annotated scenes collected by real self-driving cars. Recorded in Boston and Singapore, nuScenes features a diverse range of traffic situations, driving maneuvers, and unexpected behaviors.&lt;/p&gt;

&lt;p&gt;The dataset includes:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Full sensor suite: 32-beam LiDAR, 6 cameras and radars with complete 360° coverage&lt;/li&gt;
&lt;li&gt;1000 urban street scenes, 20 seconds each&lt;/li&gt;
&lt;li&gt;1,440,000 camera images&lt;/li&gt;
&lt;li&gt;23 classes and 8 attributes&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Sjwy6zii--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/qfchggtpoc764kqfru7m.jpeg" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Sjwy6zii--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/qfchggtpoc764kqfru7m.jpeg" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Accessing nuScenes data in SiaSearch
&lt;/h2&gt;

&lt;p&gt;To access the data yourself, you’ll need to &lt;a href="https://public.sia-search.com/"&gt;sign up for a free account on SiaSearch&lt;/a&gt;. After creating an account, you can immediately load and visualize nuScenes data within the web-based GUI.&lt;/p&gt;

&lt;p&gt;&lt;strong&gt;Now that we’re all set up, let’s start exploring!&lt;/strong&gt; Upon loading nuScenes in SiaSearch, you’ll see the main dashboard, which captures key features of the dataset. This view lets you quickly understand the overall dataset composition, as well as identify any gaps in data distribution.&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--M74C5oHr--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/7107q5y1rgbevtpn23pc.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--M74C5oHr--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/7107q5y1rgbevtpn23pc.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Querying the nuScenes Dataset
&lt;/h2&gt;

&lt;p&gt;Having a holistic view of the dataset, while useful, is not enough. The ability to drill into specific subsets can uncover insights and imbalances in the data—a critical step in model building and validation. &lt;/p&gt;

&lt;p&gt;SiaSearch makes every piece of nuScenes data searchable against all available and auto-extracted dimensions using its intelligent search interface. The platform features two ways to search for the exact sequences you want, using either a visual or code interface. The visual query lets you select extractors from a list of semantic attributes and driving situations, while the code query functions like any API call statement would.&lt;/p&gt;

&lt;p&gt;For example, if we are searching for rainy weather scenes with over three cars in view, you could submit the following query using the code interface:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;vehicle_following = 'True' AND precip_type = 'RAIN' AND 
road_joint = 'True' AND forward_velocity &amp;gt;= 5 AND num_cars &amp;gt;= 3
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Within seconds, the API returns clips that match the query attributes. This makes it easy to search the entire data lake to return situations that fit the case in question. From the search interface, specific clips can be selected by clicking into them to reveal more details, such as charts for forward velocity and acceleration profiles.&lt;/p&gt;

&lt;h2&gt;
  
  
  Curating custom training datasets from nuScenes
&lt;/h2&gt;

&lt;p&gt;Another key benefit of using the platform is the ability to transform data to produce custom training datasets. The flexibility of the SiaSearch platform allows you to finetune data for training and testing to see a measurable improvement in your models.&lt;/p&gt;

&lt;p&gt;For example, you may want to isolate the data right before or after an unprotected left turn in order to contextualize vehicle behavior. The SiaSearch interface gives you the ability to adjust the length of snippets to match your requirements in just a few clicks.&lt;/p&gt;

&lt;p&gt;Here are just a few ways you can use the SiaSearch to transform data:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Re-adjust clip length by clicking on the adjust clip button in the replay viewer. Easily trim or expand snippets according to your needs.&lt;/li&gt;
&lt;li&gt;Add custom tags to selected segments to produce new subsets of data.&lt;/li&gt;
&lt;li&gt;Leave comments for collaborators on a frame or sequence level to notify them of any questions or feedback.&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--rx8G5dpe--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5nmjm5oq0v9e9i9qoh7o.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--rx8G5dpe--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5nmjm5oq0v9e9i9qoh7o.png" alt="Alt Text"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;Now that we’ve selected our training samples, we need to find a way to feed the data back into the computer vision model.&lt;/p&gt;

&lt;h2&gt;
  
  
  Exporting data from SiaSearch
&lt;/h2&gt;

&lt;p&gt;The final step is to export your data into a format you can import directly into your model. To access the export page, add one or more snippets to your export list from either the search or playback pages. &lt;/p&gt;

&lt;p&gt;On the export page, you can review a breakdown of the selected snippets based on their associated queries, as well as basic statistics to get an idea of the quantity and contents of the export. When you’re happy with your results, you can export the raw data and or specific metadata attributes as a parquet file to feed back into your model or validation pipeline.&lt;/p&gt;




&lt;p&gt;Carefully curated data is critical in the development of any great computer vision model. At SiaSearch, we are determined to accelerate ADAS development by empowering engineers with full access and visibility into their data.&lt;/p&gt;

&lt;p&gt;&lt;em&gt;If you found this article interesting, please leave a comment and follow us for similar content. For more information about the SiaSearch platform, please check out our website at &lt;a href="https://www.siasearch.io/"&gt;https://www.siasearch.io/&lt;/a&gt;.&lt;/em&gt;&lt;/p&gt;

&lt;p&gt;&lt;em&gt;nuScenes is available for commercial use under our commercial license agreement. Please reach out to &lt;a href="mailto:nuScenes@motional.com"&gt;nuScenes@motional.com&lt;/a&gt; to learn more.&lt;/em&gt;&lt;/p&gt;

</description>
      <category>computerscience</category>
      <category>datascience</category>
      <category>machinelearning</category>
      <category>aws</category>
    </item>
    <item>
      <title>Data Curation, without the effort</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Tue, 06 Apr 2021 01:27:12 +0000</pubDate>
      <link>https://forem.com/siasearch/data-curation-without-the-effort-24a7</link>
      <guid>https://forem.com/siasearch/data-curation-without-the-effort-24a7</guid>
      <description>&lt;p&gt;When building machine learning (ML) models for highly complex automation tasks it is not only essential to identify the right models for the right job but also to select the suitable datasets for training, testing and validation. While dataset selection is typically easier for ML problems based on structured data, there are massive challenges to use data-driven technologies with unstructured data. Typical examples of such applications are robotics, automated driving and computer vision. While the content of structured data can be easily accessed and queried, dealing with unstructured data typically requires immense amounts of manual work for data selection, annotation and dataset balancing. Even tasks that seem simple, like better understanding the content of whole datasets, become extremely challenging as the typical information depth is very low as little to no additional knowledge about the content and context of the underlying unstructured data is present.&lt;/p&gt;

&lt;p&gt;Data selection is a crucial factor for all data-driven applications. It is of great importance to ensure that no biases are introduced by the data, to avoid correlation between the training and test data, and to cover the whole spectrum of possible input data as models are typically bad at extrapolating. This requires deep insights into the data which are typically hard to get, especially on a large scale (several 1000s of hours of recordings). Handling the data becomes tricky and manual labor cannot provide the necessary insights at this scale. &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--mDoy0ic7--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/wwqd9dlv8ufbv4ai8jxm.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--mDoy0ic7--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/wwqd9dlv8ufbv4ai8jxm.png" alt="SiaSearch API"&gt;&lt;/a&gt;&lt;br&gt;
&lt;/p&gt;
&lt;center&gt;&lt;em&gt;The SiaSearch API allows the user to easily access and analyze the data through efficient queries. Once the user has identified the desired subset of data, it can easily be exported and used for other workflows, like data annotation or model training.&lt;/em&gt;&lt;/center&gt;
&lt;br&gt;

&lt;p&gt;With &lt;a href="https://www.siasearch.io/product/"&gt;SiaSearch&lt;/a&gt; we provide the tools that enable users to identify the desired data within seconds. As a large amount of content information and metadata is automatically extracted during data ingestion, the data is easily searchable for the end user -- even on a petabyte scale of data.&lt;/p&gt;
&lt;h2&gt;
  
  
  Using the SiaSearch API to balance datasets
&lt;/h2&gt;

&lt;p&gt;In order to illustrate how to work with the SiaSearch API and SDK, we will walk you through the simple process of identifying data for model training and testing, for example. Let’s assume for now that the AV engineers identified problems in the object detection module while conducting lane changes. Therefore they want to improve their models to address the problem and investigate if new data can be added to the existing training and test sets. In this case, SiaSearch can be used to crawl through previously unlabeled data and identify additional lane changes.&lt;/p&gt;

&lt;p&gt;We will illustrate the following functionalities in this post:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Querying the large underlying dataset for specific situations&lt;/li&gt;
&lt;li&gt;Conducting a visual sanity check of the data&lt;/li&gt;
&lt;li&gt;Generating subsets of data and comparing the underlying data content&lt;/li&gt;
&lt;li&gt;Exporting data to a common format&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;In the following we’ll provide a few steps on how to work with our API and the Python SDK. Please note that only an excerpt of the functionalities can be shown in this post. If you are interested in obtaining full access to SiaSearch and its API, please reach out to us at &lt;a href="mailto:hi@siasearch.com"&gt;hi@siasearch.com&lt;/a&gt; or &lt;a href="https://www.siasearch.io/request-a-demo"&gt;request a demo here&lt;/a&gt;.&lt;/p&gt;
&lt;h2&gt;
  
  
  Authenticate with the API
&lt;/h2&gt;

&lt;p&gt;The first step is to use the login credentials to log into SiaSearch.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; from siasearch import auth
&amp;gt;&amp;gt;&amp;gt; sia = auth(username='foo', password='bar', "https://some-endpoint.com")
&amp;gt;&amp;gt;&amp;gt; sia
Siasearch object with user `foo` connected to `https://some-endpoint.com
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Querying for desired data
&lt;/h2&gt;

&lt;p&gt;Queries can be constructed with an SQL-like DSL (domain specific language). For our simple example where we want to find lane changes in our own “mxa” dataset, the query looks like this:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; results = sia.query("dataset_name = 'mxa' AND is_lane_change = 'True'")
&amp;gt;&amp;gt;&amp;gt; len(results.segments)
56
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;The API returns a results object which consists of segments that match the query attributes. Those segments are defined by the dataset, the underlying drive IDs and the start and end time of each segment and have various methods relating to how a user can interact with a segment. &lt;/p&gt;

&lt;p&gt;The results can also be obtained as a Pandas DataFrame.&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; results.df_segments.head()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--rmGfT09---/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/j763eajjf7kvpu17qttk.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--rmGfT09---/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/j763eajjf7kvpu17qttk.png" alt="Pandas DataFrame"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;h2&gt;
  
  
  Validation of results
&lt;/h2&gt;

&lt;p&gt;It's critical to validate the query results. The query we previously constructed may have been too narrow, or vague. In either case, we can iterate and tweak the query until we are satisfied with the resulting set of events.&lt;/p&gt;

&lt;p&gt;The easiest way to visually inspect the query results is by accessing them in the SiaSearch web interface. Giving you an overview of all segments, and allowing you to inspect a single segment in more detail. &lt;/p&gt;

&lt;p&gt;This can be achieved with the following command:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; results.url()
https://demo.sia-search.com/search?query=dataset_name+=+%27mxa%27+AND+is_lane_change+=+%27True%27&amp;amp;submit=true
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;This then provides you with an overview of the segments and a statistical aggregation of the results:&lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--YMo-2cNw--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/bbr8l6byqbc8qgkmmvu0.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--YMo-2cNw--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/bbr8l6byqbc8qgkmmvu0.png" alt="image"&gt;&lt;/a&gt;&lt;br&gt;
&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--K9lANlly--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/1lnk3el1emmvkj50xbiw.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--K9lANlly--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/1lnk3el1emmvkj50xbiw.png" alt="image"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The GUI also allows the user to dig deeper into each segment and use the embedded data viewer: &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--aZB4wyiQ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/bjaffuydnq1vu7b1laz7.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--aZB4wyiQ--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/bjaffuydnq1vu7b1laz7.png" alt="image"&gt;&lt;/a&gt;&lt;/p&gt;
&lt;h2&gt;
  
  
  Generating data subsets by tagging
&lt;/h2&gt;

&lt;p&gt;In our simple example the goal was to identify unseen lane changes which are potentially interesting to use for model training and testing. Therefore we also want to be able to tag segments and produce subsets of data. Here we will show you how to tag the first two-thirds of the segments as training data and the rest as test data:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; switching_point = int(len(results.segments) *2/3)
&amp;gt;&amp;gt;&amp;gt; for seg in results.segments[:switching_point]:
        seg.add_tag("lane_change_train")
&amp;gt;&amp;gt;&amp;gt; for seg in results.segments[switching_point:]:
        seg.add_tag("lane_change_test")
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;It is easy to obtain information about which tags are already present for the current user by calling:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; sia.get_all_tags()
['lane_change_test', 'lane_change_train']
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;Or simply return all segments which belong to a certain tag:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; tag_results = sia.get_results_from_tag('lane_change_test')
&amp;gt;&amp;gt;&amp;gt; tag_results.segments[0]
&amp;lt;Segment(drive_id='mxa_drive_0052', start_timestamp='2020-02-06 15:04:47.400000+00:00', , dataset_name='mxa', tag='lane_change_test')&amp;gt;
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;h2&gt;
  
  
  Accessing additional metadata
&lt;/h2&gt;

&lt;p&gt;If we want to build analytics and visualizations without depending on the SiaSearch web interface we can access SiaSearch’s automatically extracted metadata catalogue through the API. In our example we will also extract information about the vehicle velocity, the road types and the precipitation types under which the lane changes were conducted. &lt;/p&gt;

&lt;p&gt;The additional metadata for the desired dimensions can be easily accessed and the results are obtained as a dataframe:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; meta_results = results.segments[0].get_raw_values(["forward_velocity", "precip_type", "tag_road_type"])
&amp;gt;&amp;gt;&amp;gt; meta_results.head()
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--Zh015T1f--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/2fag1b6298ue275nl234.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--Zh015T1f--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/2fag1b6298ue275nl234.png" alt="image"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;This dataframe allows us to quickly generate new plots in order to analyze the underlying data. In our example this allows us to compare the train and test datasets in a straightforward way by visualizing the additional metadata: &lt;/p&gt;

&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--dPwRvP9X--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/zopq2ny5uc2vl3dzjf4j.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--dPwRvP9X--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/zopq2ny5uc2vl3dzjf4j.png" alt="image"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The underlying GPS data can also be obtained in order to generate geospatial visualizations (using ipyleaflet for jupyter notebooks).&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; m = Map(center=compute_gps_center(results), zoom=7)
&amp;gt;&amp;gt;&amp;gt; add_segments_to_map(
        m, 
        sia.get_results_from_tag('lane_change_train').segments,
        color="blue"
    )
&amp;gt;&amp;gt;&amp;gt; add_segments_to_map(
        m, 
        sia.get_results_from_tag('lane_change_test').segments, 
        color="red"
    )
&amp;gt;&amp;gt;&amp;gt; m.save("my_map.html", title="My Map")
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;&lt;a href="https://res.cloudinary.com/practicaldev/image/fetch/s--g4RjOE1l--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5cued54glhwxn9qh5f5w.png" class="article-body-image-wrapper"&gt;&lt;img src="https://res.cloudinary.com/practicaldev/image/fetch/s--g4RjOE1l--/c_limit%2Cf_auto%2Cfl_progressive%2Cq_auto%2Cw_880/https://dev-to-uploads.s3.amazonaws.com/uploads/articles/5cued54glhwxn9qh5f5w.png" alt="image"&gt;&lt;/a&gt;&lt;/p&gt;

&lt;p&gt;The visualizations obtained through the API clearly show that the data in the train and test sets are not balanced, which could cause biases during training or misleading test results, as most test data was recorded on highways. This is not surprising since the process we used to split the data is very simple (first two thirds of the data put into the training set, the rest in the test set).  More iterations or a different split of the data would be required.&lt;/p&gt;

&lt;h2&gt;
  
  
  Exporting
&lt;/h2&gt;

&lt;p&gt;The final step is to export the results into a format you can use for importing into your model or validation pipeline.&lt;/p&gt;

&lt;p&gt;The easiest way to achieve this is to access the underlying Panda’s DataFrame which provides a multitude of export formats to choose from. See the official Panda’s document for all available export formats.&lt;/p&gt;

&lt;p&gt;As an example:&lt;br&gt;
&lt;/p&gt;

&lt;div class="highlight js-code-highlight"&gt;
&lt;pre class="highlight plaintext"&gt;&lt;code&gt;&amp;gt;&amp;gt;&amp;gt; df_results = results.df_segments
&amp;gt;&amp;gt;&amp;gt; df_results.to_csv(“out.csv”)
&amp;gt;&amp;gt;&amp;gt; df_results.to_parquet(“out.parquet”)
&lt;/code&gt;&lt;/pre&gt;

&lt;/div&gt;



&lt;p&gt;In future versions we will provide more custom formats, not directly available from Pandas interface, such as GeoJSON, KML, and MAT files.&lt;/p&gt;

&lt;p&gt;As shown in the previous steps, the SiaSearch API and Python SDK allow the user to easily interact with large scale unstructured data. Cumbersome data engineering tasks are abstracted away which allows the user to focus on the important task of analyzing the data and identifying the required datasets for training, testing and validation files.&lt;/p&gt;

&lt;h2&gt;
  
  
  Get started with SiaSearch
&lt;/h2&gt;

&lt;p&gt;If you share some of the problems we talked about, or are excited give SiaSearch a try, &lt;a href="https://www.siasearch.io/request-a-demo"&gt;request a demo&lt;/a&gt; on our website or reach out to us at &lt;a href="//mailto:hi@siasearch.io"&gt;hi@siasearch.io&lt;/a&gt;.&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Originally published by Mark Pfeiffer on: &lt;a href="https://www.siasearch.io/blog/data-curation-without-the-effort"&gt;https://www.siasearch.io/blog/data-curation-without-the-effort&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>datascience</category>
      <category>computerscience</category>
    </item>
    <item>
      <title>Automotive Companies only Access 5% of their Vehicle Data</title>
      <dc:creator>SiaSearch</dc:creator>
      <pubDate>Mon, 05 Apr 2021 06:02:39 +0000</pubDate>
      <link>https://forem.com/siasearch/automotive-companies-only-access-5-of-their-vehicle-data-n71</link>
      <guid>https://forem.com/siasearch/automotive-companies-only-access-5-of-their-vehicle-data-n71</guid>
      <description>&lt;p&gt;With vehicle sensors collecting massive amounts of data, only 5% of it is currently being used for product development. Better infrastructure and data processing hold the keys to progress.&lt;/p&gt;

&lt;p&gt;The promise of fully autonomous vehicles continues to excite and inspire millions of people around the world. The amazing things that safe, reliable, self-driving vehicles can do for humanity--from providing newfound mobility to senior citizens to reducing traffic accidents--are closer to our reach than ever. But we still have a long way to go. &lt;/p&gt;

&lt;p&gt;Along with the fuel (gasoline, diesel or electricity) that powers automobiles, autonomous vehicles require a fuel of their own to “drive” safely and effectively: data. Although that data, already collected by millions of sensors on thousands of vehicles around the world, is readily available, it’s not being utilized to its full potential.&lt;/p&gt;

&lt;p&gt;These datasets power the algorithms that make all levels of autonomous driving possible. Today, automotive companies only access 5% of their vehicle data, while the remaining 95% becomes costly to store and optimise without the context necessary to make use of it. As automotive companies aggressively pursue data-driven product development, data handling, including the exploration, querying, curation and evaluation of data, is a common bottleneck on the road to progress. These essential but hard-to-manage datasets bring a unique set of challenges for those hoping to make use of them:&lt;/p&gt;

&lt;ul&gt;
&lt;li&gt;Unstructured and diverse formats &lt;/li&gt;
&lt;li&gt;Need for rich semantics in order to access them &lt;/li&gt;
&lt;li&gt;Huge sizes that require high-performance computing
&lt;/li&gt;
&lt;li&gt;Strong need for data versioning &lt;/li&gt;
&lt;li&gt;Access and security issues &lt;/li&gt;
&lt;li&gt;Need for continuous failure-case driven data exploration&lt;/li&gt;
&lt;/ul&gt;

&lt;p&gt;With &lt;a href="https://www.siasearch.io/product"&gt;proper data processing and analytics software&lt;/a&gt;, engineering users can overcome all of these challenges and vehicle data can fulfill its potential. By upgrading legacy technology for data access and analysis, OEMs and mobility tech companies can bump up data utilisation rates by up to 40% and generate additional ROI as data exploration, search, analysis, anomaly detection and evaluation require less manual engineering work and yield better results. &lt;/p&gt;

&lt;h2&gt;
  
  
  Data infrastructure and management are being neglected
&lt;/h2&gt;

&lt;p&gt;Currently, vehicle technology developers are focused on machine learning models and ground truth labeling. These same developers are neglecting infrastructure upgrades, leading many to use legacy technology for data management. Terabytes of unstructured, unprocessed vehicle data easily overwhelm these systems, causing them to malfunction. Raw data has no metadata and billions of frames, and technology developers are left to use tools ill-suited to the task of data management to organize data on their own. &lt;/p&gt;

&lt;p&gt;Highly-paid engineers search datasets manually on sluggish database systems, spending up to 75% of their time on raw data handling issues instead of training and validating building models.  Furthermore, the lack of insights garnered from these vehicle datasets means machine learning and data science systems are unable to effectively build AI functions that rely on sensor data as an input.&lt;/p&gt;

&lt;h2&gt;
  
  
  Purpose-built data infrastructure is the solution
&lt;/h2&gt;

&lt;p&gt;Infrastructure that is designed and built specifically to house raw sensor data and extract insights is the answer. This infrastructure should be as simple and easy to navigate as spreadsheets and SQL databases that bring order and usefulness to data in other industries. For vehicle sensor data, infrastructure that uses a flexible and minimalist data model and a scalable method to produce semantics, along with fast queries and integrated endpoints to use and share data, is most effective.&lt;/p&gt;

&lt;p&gt;These and other infrastructure elements improve data re-use and result in faster insights. Enabled by semantic automation, data can be ranked according to its importance based on context, content, criticality, and usability. This reduces redundancy and maximizes information density. Low importance data is archived or deleted, while high importance data is easily accessible for everyone. Furthermore, data insights can be provided almost immediately with overviews of the content and the redundancy. &lt;/p&gt;

&lt;p&gt;Users can augment retention recommendations with tailored rule-based constraints, e.g. unprotected left turns should always be kept. &lt;/p&gt;

&lt;h2&gt;
  
  
  Better autonomous vehicles, sooner
&lt;/h2&gt;

&lt;p&gt;Advances in automated data infrastructure and processing liberate automotive technology developers from the constraints of legacy technology systems. Automated, semantic data management systems make data handling easier than ever, enabling automotive companies to save valuable engineering time, boosting productivity and increasing overall utilization.   &lt;/p&gt;

&lt;p&gt;Put simply, more robust, effective infrastructure for unstructured sensor data will lead to higher ROI on research and development by freeing up engineers to do what they do best: building algorithms that will power the autonomous vehicles of the future. And with those engineers working more efficiently, the dreams of autonomous vehicles improving our lives are that much closer to reality. &lt;/p&gt;

&lt;h2&gt;
  
  
  Dealing with lots of raw sensor data?
&lt;/h2&gt;

&lt;p&gt;Learn more about how the SiaSearch data management platform can help on our website: &lt;a href="https://www.siasearch.io/"&gt;https://www.siasearch.io/&lt;/a&gt;&lt;/p&gt;




&lt;p&gt;&lt;em&gt;Originally published by Clemens Viernickel on: &lt;a href="https://www.siasearch.io/blog/raw-sensor-data-needs-infrastructure-to-be-useful/"&gt;https://www.siasearch.io/blog/raw-sensor-data-needs-infrastructure-to-be-useful/&lt;/a&gt;&lt;/em&gt;&lt;/p&gt;

</description>
      <category>machinelearning</category>
      <category>computerscience</category>
    </item>
  </channel>
</rss>
