Skip to content

Releases: roboflow/inference

v0.26.0

08 Nov 17:19
b8a5f64
Compare
Choose a tag to compare

🚀 Added

🧠 Support for fine-tuned Florence-2 💥

As part of onboarding of Florence-2 fine-tuning on Roboflow platform, @probicheaux made it possible to run your fine-tuned models in inference. Just complete the training on the Platform and deploy it using inference, as any other model we support 🤯

🚦Jetpack 6 Support

We are excited to announce the support for Jetpack 6 which will enable more flexibility of development for Nvidia Jetson devices.

Test the image with the following command on Jetson device with Jetpack 6:

pip install inference-cli

inference server start

or pull the image from

docker pull roboflow/roboflow-inference-server-jetson-6.0.0

🏗️ Changed

InferencePipeline video files FPS subsampling

We've discovered that the behaviour of max_fps parameter is not in line with inference clients expectations regarding processing of video files. Current implementation for vides waits before processing the next video frame, instead dropping the frames to modulate video FPS.

We have added a way to change this suboptimal behaviour in release v0.26.0 - new behaviour of InferencePipeline can be enabled setting environmental variable flag ENABLE_FRAME_DROP_ON_VIDEO_FILE_RATE_LIMITING=True.

❗ Breaking change planned

Please note that the new behaviour will be the default one end of Q4 2024!

See details: #779

Stay tuned for future updates!

Other changes

🔧 Fixed

🏅 New Contributors

Full Changelog: v0.25...v0.26.0

v0.25

01 Nov 18:54
727ebd0
Compare
Choose a tag to compare

🚀 Added

Newer onnxruntime 👉 faster inference on MacBook 🐎

@yeldarby bumped the onnxruntime from 0.15.x into 0.19.0 and that small change induced great performance improvement for YOLOv8 models running on resolution up to 640 🤯 Some variants of yolo runs now 2x faster on M2 chips ❗

profiling.mp4

🏎️ Better and Faster OWLv2 inference

@isaacrob-roboflow and @probicheaux prepared something special in #759, #763 and #755

🧙 Box prompting in annotations

We built a new tool to automatically label your data for you!

Inspired by a recent open source release, now when you annotate a dataset you can automatically get recommendations for other likely bounding boxes! And if you don’t like a recommendation, you can mark it as ‘negative’ and it will learn not to recommend boxes like that again!

Under the hood we use your annotations and feedback to train a few shot model based on OWLv2 on the fly, and then run it against your data to propose other likely bounding boxes.

We think this will save you a lot of time!

🐎 Inference speedup

Over 10x inference speed boost on T4 GPU. Previous version run in 440ms, now it is only 36ms 🔥

Workflows enhancements

🅰️ Single character detection 👉 OCR

Ever wondered how to read the text if you have model detecting separate characters? Now it is easy with workflows thanks to @reiffd7

Check out Stitch OCR Detections block docs 📖

ocr_stitch.mp4

🤺 Keypoints detection visualisation

Thanks to @EmilyGavrilenko, Workflows can now visualise predictions from keypoints detection models - for instance pose-estimation ones 😄 I bet you find the visualisation familiar - the new block is powered by supervision.

Check out Keypoint Visualization block docs 📖

keypoints_visualization.mp4

🔧 Fixed

🏅 New Contributors

Full Changelog: v0.24.0...v0.25.0

v0.24.0

18 Oct 17:09
4090999
Compare
Choose a tag to compare

🚀 Added

🎥 Data analysis and export in Workflows

We’re excited to introduce a suite of new blocks to supercharge your video processing and streamline your workflow exports!

✨ What’s New?

  • Enhanced Video Analytics: Easily track object counts over time or within specific zones. Want to know how many items appear in a unit of time? ✅ Now you can!
  • Custom Notifications: Send automatic emails 📧 when a detected object enters a defined area. Stay informed without manual checks!
  • REST API Integration: Effortlessly export your workflow results 🌐. Connect to any REST API and deliver your data wherever you need it.

Our new blocks make your Workflows more powerful and versatile than ever before. Start building smarter workflows today! 🚀

video_analysis_with_workflows.mp4

New blocks

  • Data Aggregator: Collects and processes Workflow data to create time-based analytics. 📊 Supports custom aggregation strategies, making it easy to summarize data streams efficiently.
  • CSV Formatter: Formats the data into CSV files which can be saved or send as an attachment to in notification block 📚
  • Email Notification: Send email notifications 📧 in Workflows.
  • Local File Sink: Saves data generated in Workflow runtime into local file 📁
  • Webhook SInk: Enables users to integrate their Workflows with REST API to export data or analysis results

Important

At the moment of release, Workflows UI lack some capabilities to display new blocks, but stay tuned - we will fix that shortly

🏷️ Tracking stabiliser block

Ever experienced flickering detections and loosing tracker id? We know this pain - that's why @grzegorz-roboflow prepared Detections Stabilizer block.

376666040-4a259335-ac24-4af0-a1c5-46b4bc27f093.1.mp4

💻 Improvements in stream processing

Each week we are closer and closer to enabling full-blown video processing features - this week @grzegorz-roboflow and @hansent pushed us forward:

  • WebRTC streams are now faster - #751
  • we have better management of InferencePipelines in inference server - #751

🔧 Fixed

🌱 Changed

  • Adding more inference metrics to the prometheus scraper by @bigbitbus in #750

Full Changelog: v0.23.0...v0.24.0

v0.23.0

11 Oct 15:24
d7879c9
Compare
Choose a tag to compare

🚀 Added

🕵️ Predictions difference visualization in Workflows 🔎

Ever felt like a detective trying to uncover the subtle differences between model predictions? It used to be hard—sifting through data, struggling to spot where models disagreed. But not anymore! 🔦

The Model Comparison Visualization block makes comparing models effortless by visually highlighting the differences between two models’ predictions.

Don't you think that was a great idea by @yeldarby to create the block?

✨ Just take a look how it works

multiple_models_comparison.mp4

🐎 Compilation cache in Workflows Execution Engine 🚀

If you're familiar with programming, you know the trade-off between compiled and interpreted languages: interpreted ones offer flexibility, but interpreting every line of code can slow things down. To overcome this, interpreters often cache code to speed up execution.

We realized that the Workflows Execution engine was adding unnecessary graph interpretation overhead, even when running the same workflow multiple times. 🕒 In the quest for speed, we’ve introduced compiled Workflow caching, which drastically reduces execution time for repeated workflows! ⚡

When you will be able to see the speed-up? If you send multiple requests to inference server asking to run the same Workflow - you will be impacted 😄 Now your workflows will run faster than ever! 🎯

Full details and benchmark results are available here

📐 Distance measurement in Workflows 📏

Introducing the Distance Measurement block, now available in Workflows! This block calculates the distance between two bounding boxes on a 2D plane using a camera positioned perpendicularly.

🔍 Key Features:

  • Measure distance in centimeters or pixels with ease.
  • Use either a reference object of known size or a pixel-to-centimeter ratio for accurate scaling.
  • Whether you're analyzing footage or comparing object placements, this block gives you precise, scalable distance measurements in your workflows. Try it out and streamline your 2D distance calculations! 📐

Everything thanks to @ediardo 🏅

📊 inference server telemetry

Thanks to @robiscoding, inference server gained ability to export detailed metrics that can be pooled in runtime to ensure better service observability. Check out details.

🗞️ Exciting news

🎥 Video processing with Workflows

The Roboflow team is pushing the boundaries of video processing within Workflows, laying the foundation for powerful new capabilities. You can catch a glimpse of the preliminary changes we've introduced thanks to @grzegorz-roboflow: #719, #743,

Stay tuned as we continue to build and enhance video processing features, making it easier than ever to integrate video into your workflows. 📹✨

🌱 Changed

Deprecation of video_metadata kind

The video_metadata kind has been deprecated. We recommend switching to the updated image kind, which now supports optional video_metadata. We plan to remove video_metadata kind in Execution Engine v2.0.0

While this is non-breaking, some older blocks may become incompatible with future video processing blocks.

With the updated internal representation of image kind, two new methods, WorkflowImageData.copy_and_replace(...) and WorkflowImageData.create_crop(...), have been added for easier metadata manipulation. See the updated usage guide.

From now one, we will also publish list of changes in Workflows Execution Engine in a changelog 📖

Remaining changes

🏅 New Contributors

Full Changelog: v0.22.2...v0.23.0

v0.22.2

04 Oct 19:13
18c9f90
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.22.1...v0.22.2

v0.22.1

04 Oct 16:02
d852dea
Compare
Choose a tag to compare

Bumps ONNX Runtime requirements to ensure our users can run the new yolov11 models. Also adds some yolov11 documentation to the aliases page.

What's Changed

Full Changelog: v0.22.0...v0.22.1

v0.22.0

04 Oct 14:33
c70cc32
Compare
Choose a tag to compare

🚀 Added

🔥 YOLOv11 in inference 🔥

We’re excited to announce that YOLOv11 has been added to inference! 🚀 You can now use both inference and the inference server to get predictions from the latest YOLOv11 model. 🔥

All thanks to @probicheaux and @SolomonLake 🏅

skateboard_yolov11.mov
Try the model ininference Python package
import cv2
from inference import get_model

image = cv2.imread("<your-image>")
model = get_model("yolov11n-640")
predictions = model.infer(image)

print(predictions)

💪 Workflows update

Google Vision OCR in workflows

Thanks to open source contribution from @brunopicinin we have Google Vision OCR integrated into workflow ecosystem. Great to see open source community contribution 🏅

google_vision_ocr.mp4

See 📖 documentation of the new block to explore it's capabilities.

Images stitch Workflow block

📷 Your camera is not able to cover the whole area you want to observe? Don't worry! @grzegorz-roboflow just added the Workflow block which would be able to combine the POV of multiple cameras into a single image that can be further processed in your Workflow.

image 1image 2stitched image

📏 Size measurement block

Thanks to @chandlersupple, we can now measure actual size of objects with Workflows! Take a look at 📖 documentation to discover how the block works.

image

Workflows profiler and Execution Engine speedup 🏇

We've added Workflows Profiler - ecosystem extension to profile the execution of your workflow. It works for inference server requests (both self-hosted and on Roboflow platform) as well as for InferencePipeline.

image

The cool thing about profiler is that it is compatible with chrome://tracing - so you can easily grab profiler output and render it in Google Chrome browser.

To profile your Workflow execution use the following code snippet - traces are saved in ./inference_profiling directory by default.

from inference_sdk import InferenceHTTPClient

client = InferenceHTTPClient(
    api_url="https://detect.roboflow.com",
    api_key="<YOUR-API-KEY>"
)
results = client.run_workflow(
    workspace_name="<your-workspace>",
    workflow_id="<your-workflow-id>",
    images={
        "image": "<YOUR-IMAGE>",
    },
    enable_profiling=True,
)

See detailed report regarding speed optimisations in the PR #710

❗ Important note

As part of speed optimisation we enabled server-side caching for workflows definitions saved on Roboflow Platform - if you frequently change and your Workflow, to see results immediately you need to specify use_cache=False parameter of client.run_workflow(...) method

🔧 Fixed

🌱 Changed

🏅 New Contributors

We do want to honor @brunopicinin who made their first contribution to inference in #709 as a part of Hacktoberfest 2024. We invite other open-source community members to contribute 😄

Full Changelog: v0.21.1...v0.22.0

v0.21.1

30 Sep 18:17
dc9362b
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.21.0...v0.21.1

v0.21.0

27 Sep 14:02
23dd6c9
Compare
Choose a tag to compare

🚀 Added

👩‍🎨 Become an artist with Workflows 👨‍🎨

Ever wanted to be an artist but felt like you lacked the skills? No worries! We’ve just added the StabilityAI Inpainting block to the Workflows ecosystem. Now, you can effortlessly add whatever you envision into your images! 🌟🖼️

Credits to @Fafruch for origin idea 💪

inpainting_stability_ai_demo.mp4
📖 docs

🤯 Workflows + video + inference server - Experimental feature preview 🔬

Imagine creating a Workflow in our UI, tuning it to understand what happens in your video. So far, video processing with InferencePipeline required a bit of glue code and setup in your environment. We hope that soon, you won’t need any custom scripts for video processing! You’ll be able to ship your Workflow directly to the inference server, just pointing which video source to process.

We're thrilled to announce that we’ve taken the first step toward making this idea a reality! Check out our experimental feature for video processing, now controlled by the inference server with a user-friendly REST API for easy integration.

video_processing_behind_api.mp4

🔍 We encourage you to try it out! The feature is available in the inference server Docker images that you can self-host. Please note that this feature is experimental, and breaking changes are to be expected. Check out our 📖 docs to learn more.

🙃 Flips, Rotations and Resizing in Workflows

Tired of dealing with image orientation problems while building demos with Workflows? Whether it's resizing, rotating, or flipping, those headaches end today with our new features for seamless image adjustments!

All thanks to @EmilyGavrilenko and PR #683

✨ Ensure Your Tracked Objects Stay on Course! 🛰️

Wondering if the objects you're tracking follow the path you intended? We’ve got you covered in Workflows! Thanks to @shantanubala, we now offer Fréchet Distance Analysis as a Workflow block. Simply specify the desired trajectory, and Workflow calculates the deviation for each tracked box. 📊
See details: #682

What’s Fréchet Distance?
It’s a mathematical measure that compares the similarity between two curves—perfect for analyzing how closely your tracked objects follow the path you’ve set.

🆕 Background removal in Dynamic Crop and updated UI for VLMs

Let’s be honest—VLMs in Workflows still had room for improvement, especially when integrating model outputs with other blocks. Well, we've made it better! 🎉 Now, each model task comes with a clear description and a suggested parser to follow the block, helping you get the most out of your model predictions with ease. 🛠️

Additionally, you can now remove background while performing Dynamic Crop on Instance Segmentation model results 🤯

CleanShot.2024-09-26.at.18.38.33.mp4

🔧 Fixed

🌱 Changed

🏅 New Contributors

Full Changelog: v0.20.1...v0.21.0

v0.20.1

24 Sep 15:32
658cb3f
Compare
Choose a tag to compare

What's Changed

Full Changelog: v0.20.0...v0.20.1