Skip to content

Multithreading inference in Tensorflow Estimators. This is a ServiceNow Research project that was started at Element AI.

License

Notifications You must be signed in to change notification settings

ServiceNow/multithreaded-estimators

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

ServiceNow completed its acquisition of Element AI on January 8, 2021. All references to Element AI in the materials that are part of this project should refer to ServiceNow.

Multithreaded-estimators

Code demonstrating how to use multithreading to speedup inference for Tensorflow estimators.

Installation

A Dockerfile is provided. First build the image from the root directory:

docker build . -t threaded

Then run the tests:

docker run threaded

License

This code is released under an Apache 2 license. See the license in full.

About

Multithreading inference in Tensorflow Estimators. This is a ServiceNow Research project that was started at Element AI.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 2

  •  
  •