Release notes

2.5.1

October 18, 2023

Minor updates

  • Added the ability to contact a human specialist through Bel Esprit.
  • Added the ability to copy Bel Esprit session ID and export the chat log.
  • Added asset renaming and editing meta information.
  • Created pipeline drafts listing page with function and visibility filters.

SDK updates

SDK build 0.2.4

  • Model image upload SDK.

Bug fixes

  • Fixed an issue with the asset drawer where it would clear assets after being used.

2.5.0

September 21, 2023

Major updates

  • Bel Esprit — We are thrilled to introduce Bel Esprit, your personal AI solution architect. Converse with an AI chat agent in natural language to transform your ideas into deployable, production-ready AI solutions.
  • Pipeline drafts in Design — A new state for pipelines in Design has been added which allows the user to save unfinished Design pipelines as drafts. Users can load drafts and pick up their work where they left off. Drafts also work with auto-saving to ensure that no work gets lost.
  • Pipeline templates in Design — We are adding pipeline templates into Design. You can now save your pipelines as templates, you can then reuse them or load pre-existing templates to speed up your pipeline building process.

Minor updates

  • Improved the documents page in the side panel and added several new articles.
  • Added the ability for users to sign-up or sign-in using their Google accounts.
  • Introduced conversation history in “Try it out” for LLMs.
  • Increased the limit for the number of access keys per team to 10.
  • Disabled the zoom button if the zoom level is already at maximum or minimum in Design.

SDK updates

  • Added data validations for the format constraints while uploading datasets.
  • Added the ability to filter models by AI function when listing them.
  • Fixed FineTune and data asset functional tests.
  • Fixed metric example displayed in the documentation.

2.4.2

September 4, 2023

Minor updates

  • Added the ability for decision node in Design to handle multiple inputs and multiple outputs.
  • Improved the Model comparison feature to handle any function with multi-input support.
  • Added Models and Metrics display in Pipeline specification table.
  • Added the ability to edit owned files and Pipelines through asset editor.
  • “Try it out” feature now handles Models with Search function.
  • Added an aiXplain Credit purchase interaction that’s triggered when the user’s balance is low.
  • Added data format requirements to Dataset and Corpus creation.

SDK updates

  • Made pipeline logs accessible to users via the SDK.

Bug fixes

  • Fixed an issue where the list of models was not updated when a model is unsubscribed.
  • Fixed an issue with handling label inputs in “Try it out”.

2.4.1

August 10, 2023

Major updates

  • FineTune for Search — We are thrilled to introduce the latest feature added to FineTune, FineTune for Search. This feature allows you to customize your own multi-modal search engine using your own data.
  • New Model and Pipeline tryout — We expanded the capabilities of “Try it out!” which is used to run and test models and pipelines on aiXplain. The new update takes a conversational approach and supports models and pipelines with multi-input, multi-output, and multi-modality.

Minor updates

  • Added the ability to download pipeline logs when you try out a pipeline.
  • Added a feedback form which is accessible from the left panel.
  • Added an aiXplain Credit purchase interaction when the user’s balance is low for “Try it out!”.
  • Added the ability to delete assets from asset cards.

SDK updates

  • Refactoring Benchmark on the SDK — We refactored Benchmark SDK to make it consistent with the previously released FineTune SDK.
  • Onboard Dataset and Corpus from an s3 bucket — We added the ability to onboard datasets and corpora to aiXplain from public s3 buckets.

Bug fixes

  • Fixed an issue with the asset drawer compatibility which allowed collecting assets that are not onboarded yet.

2.4.0

July 24, 2023

Major updates

  • Metric changes — In this release, we wanted to increase the versatility of how evaluation metrics that are used in Benchmark are behaving. Hence, we are adding more details to Metric cards to be more descriptive. Adding details like the input, output, pricing, and supplier of each Metric. This would also give more visibility to the pricing of Benchmark jobs.
  • Text normalization setting in Benchmark — A new feature has been added to Benchmark which allows using various normalization settings for text to be used. This feature comes with minor UI improvements to Benchmark job configuration.

Minor updates

  • Added the ability to use public URLs as valid data in Dataset and Corpus upload.
  • Added the ability to filter models in Discover which are FineTune compatible.
  • Added tooltips for different types of nodes in Design when you hover over them.
  • Added double-click functionality in Design to re-align zoom and access node details.
  • Improved UI visuals of Design nodes with unified asset node types and dynamic font size.
  • Added autosave functionality for pipelines in Design.
  • Added a Dashboard guide that is visible for the first time a member signs.

SDK updates

  • SDK for FineTune — SDK now supports fine-tuning models.

Bug fixes

  • Fixed an issue with data persistence that kept experience progress when switching teams.
  • Fixed an issue with handling third party packages which impacted certain nodes in Design.

2.3.0

June 12, 2023

Major updates

  • Corpus and Dataset changes — In release 2.2.0 we introduced a new asset type “Corpus” which allowed structured data to be onboarded on aiXplain without having to allocate a function to it.
    • Added the ability to create Datasets from Corpora, or use them directly with your favorite tools like Benchmark and FineTune.
    • Improved Dataset creation process to allow for more options such as adding multiple references and handling errors.
    • Added Datasets and Corpora creation from files already uploaded to aiXplain instead of only supporting local files.
    • Added Dataset and Corpus logs which are available on the details page of the assets to show the history of these assets.
  • AutoMode training temporarily disabled — In light of the changes to datasets in aiXplain, our team is working hard to address their compatibility with AutoMode. This won’t affect the already trained AutoMode models. Expect an update that will address this soon.

Minor updates

  • Added more information on the Corpus details page that shows the number of columns/features and rows/segments in a Corpus.
  • Added credit expiration notice for Credits granted through phone verification.
  • Updated the colors in SHAP plots inside Benchmark reports.
  • Added more metadata categories in Benchmark reports.

Bug fixes

  • Fixed a bug that showed the “Compare” button in the asset drawer for assets that don’t support this function.
  • Fixed a bug that displayed the wrong labels in nodes inside Design.
  • Fixed an issue with certain models that failed when rerunning a Benchmark job.

2.2.0

April 26, 2023

Major updates

  • Discover changes — In preparation for bigger upcoming changes to some of our AI asset types, we updated the design of Discover to make finding the assets that you are looking for easier and more intuitive! The changes will also be reflected on your assets that are accessible from your Dashboard.
  • Introducing a new asset type: Corpus — In this release we added a new asset type that you can find, onboard, and use in aiXplain. The corpus is structured data which is not defined to be used for a single particular AI function. There are limited things that you can do with corpora for the time being but our next release will allow extracting multiple datasets for multiple AI functions from a single corpus. For example, I can onboard a corpus with columns for speech, transcript, translation, speaker age, speaker gender, speaker ID where I can use it to create datasets for Speech Recognition, Speaker Diarization, Speech Classification, and Translation.
  • Updates to API access keys — We previously added the ability for teams to generate access keys from the team settings. In this release, we removed the auto-generated default access key which was specific to a particular model or pipeline for each team. This change allows creating and using different access keys for different endpoints. With added security allowing members to delete and recreate access keys as they see fit.

Minor updates

  • Translation metrics added in Design — We added all of the Translation metrics found in Benchmark to the metric node inside Design.
  • Improved Benchmark — We improved the way that scoring in Benchmark takes place. Those changes will have positive impact on Benchmark performance and stability.
  • Enhanced network security — We added improved encryption to the tokens sent out over the network.
  • User guide in Dashboard — Added a user guide in the Dashboard to help the members access the documents page and use the platform.

Bug fixes

  • Fixed a bug that made Benchmark reports appear to be stuck at 90% when they are complete.
  • Fixed a bug where members couldn’t find the “switch team” option when the team invitation is accepted.
  • Fixed a bug that prevented the phone number verification to appear on the dashboard.
  • Fixed a bug in displaying metrics inside the metric node in Design that collapsed the scroll.
  • Fixed a bug where some function names were displayed incorrectly.

2.1.1

February 8, 2023

Minor updates

  • Phone verification — Go to your account settings and verify your phone number to get access to your free credits!
  • Renaming pipelines — Added the ability to rename pipelines after creating them.

Bug fixes

  • Fixed a bug where clicking on canvas doesn’t close dropdown menus in Design.
  • Fixed a bug where the back button wasn’t working as intended in some areas of the app.
  • Fixed a bug where you were redirected to the wrong URL after payment method is updated.

2.1.0

January 12, 2023

Major updates

  • FineTune tool — We have launched the ability to fine-tune select models for MT and ASR. You can also view the logs of the model inside the details page for it. Read more
  • Segmentor node and constructor node — We added new nodes into Design! Using the segmentor node and constructor node, members now have the ability to customize the segmentation of the data flowing through their pipelines and reconstruct them back. This optimizes the use of Design for models where the input size matters.
  • Metric node — We added another new node in Design! Members can add benchmarking to their pipelines as a node which gives them the ability to evaluate the performance of their models while a pipeline is running.
  • Activity logs — We added a log for all your activities on aiXplain which allows you to filter them by product and by status. This will give members more control and insight over what activities are being done on their teams such as model fine-tuning and Benchmark reports.

Minor updates

  • Imputation of failed inputs — Now in Benchmark, members have the option to impute the scores of some failed segments to get a higher representation of their data.
  • Added the number of segments to be displayed onto the asset card for datasets.
  • Added the ability to create API keys on aiXplain: In the top menu under team settings, members will be able to generate API keys which can then be used as a universal key to run all of their assets.
  • Removed phone number from registration.
  • Added icons next to AutoMode and FineTune models.
  • Added multiple notifications in platform for Benchmark, FineTune, and AutoMode.

Bug fixes

  • Fixed an issue which didn’t allow saving pipelines without setting decision nodes.
  • Fixed a bug that showed a redundant message when not having enough credits on the platform.
  • Fixed a bug that allowed users to add multiple datasets in benchmarking when they shouldn’t.
  • Fixed a bug where submetrics showed redundant metrics name.
    Fixed an issue with the IQR filter displaying inside the segments tab in a benchmark report.
  • Fixed a bug where the label of the decision node couldn’t be updated.
  • Improved error handling for model and pipeline tryout.

2.0.0

November 17, 2022

Major updates

  • The new aiXplain release offers improved system performance and stability with a brand new user-friendly and intuitive design!
  • Dashboard — The aiXplain Dashboard is the go-to place for all assets from aiXplain tools.
  • Asset drawer — The asset drawer allows the collection of assets in one place where they can be used with any tool on aiXplain.
  • Derivative data — Create derivative data through pipelines or Benchmarking reports and view data history.
  • Billing nested view — This new view allows for a simpler report of the transactions made on aiXplain at a glance.
  • AutoMode — aiXplain’s AutoMode is an ensemble model that routes the input to the most optimal system according to the quality preference that it is trained on it on. The supported functions for AutoMode are Automatic Speech Recognition and Machine Translation.
  • Multi-input/output support — Design now supports connecting and running multiple input and output nodes in a single pipeline.
  • Subtitling node — The Subtitling node allows having an entire subtitling system in one node.
  • Decision node — The decision node allows members to set routing of the data in a pipeline based on set conditions and values.
  • Benchmark for Diacritization — Benchmark now supports Arabic text diacritization.
  • Benchmark for text classification — Benchmark now supports text classification.

Minor updates

  • Categories in dataset creation — Members can now specify categories when they are creating a dataset on aiXplain; these categories will be used when calculating bias in Benchmark jobs.
  • Light/dark mode — aiXplain now supports light and dark modes!
  • Downloading results as CSV — If a dataset license is owned, members can now download the results of a Benchmark report as a csv. While a Benchmark job is running, you can download the up-until-this-moment-computed results as a csv.
  • Zooming in benchmarking report plots — Now members can zoom in the benchmarking report plots.
  • Canceling and rerunning models in benchmarking — Now you can cancel and/or rerun a single model in benchmarking. This is useful for when a model fails during benchmarking.
  • Data sampling — Members can now specify how the results for Benchmark jobs can be displayed based on the data samples they’d like to use.
    • All segments — Shows the results of all the segments in a Benchmark job with a penalty score for failed segments.
    • Successful segments — Shows the results only based on the successful segments in a Benchmark job, disregarding failed segments in calculating the scores.
    • Intersecting successful segments — Shows the results only based on the common successful segments between all models.
  • Interquartile range multiplier — This configuration allows members to specify the multiplier at which the length of the whiskers in box plots is set.