Placeholder.__format__()
is now disallowed, so you cannot use placeholders in f-strings andstr.format()
calls anymore. If you get an error from this, most likely you discovered a bug and should not use an f-string in the first place. If it is truly your intention to print the placeholder (not its resolved value) for debugging purposes, userepr()
or!r
instead.
Package Name | Version Constraints | Previously (in v1.15.1 ) |
Comments |
---|---|---|---|
docker |
>=7,<8 |
>=4.1,<5 |
- Support KFP pipeline spec 2.1.0 version schema and YAML files with KFP v2 DAG runner
Package Name | Version Constraints | Previously (in v1.15.0 ) |
Comments |
---|---|---|---|
kfp-pipeline-spec |
>0.1.13,<0.2 |
>=0.1.10,<0.2 |
- Dropped python 3.8 support.
- Dropped experimental TFX Centralized Kubernetes Orchestrator
- Extend GetPipelineRunExecutions, GetPipelineRunArtifacts APIs to support filtering by execution create_time, type.
- ExampleValidator and DistributionValidator now support anomalies alert generation. Users can use their own toolkits to extract and process the alerts from the execution parameter.
- Allow DistributionValidator baseStatistics input channel artifacts to be empty for cold start of data validation.
ph.make_proto()
allows constructing proto-valued placeholders, e.g. for larger config protos fed to a component.ph.join_path()
is likeos.path.join()
but for placeholders.- Support passing in
experimental_debug_stripper
into the Transform pipeline runner.
Placeholder
and all subclasses have been moved to other modules, their structure has been changed and they're now immutable. Most users won't care (the main public-facing API is unchanged and behaves the same way). If you do special operations likeisinstance()
or some kind of custom serialization on placeholders, you will have to update your code.placeholder.Placeholder.traverse()
now returns more items than before, namely also placeholder operators like_ConcatOperator
(which is the implementation of Python's+
operator).- The
placeholder.RuntimeInfoKey
enumeration was removed. Just hard-code the appropriate string values in your code, and reference the newLiteral
typeplaceholder.RuntimeInfoKeys
if you want to ensure correctness. - Arguments to
@component
must now be passed as kwargs and its return type is changed from being aType
to just being a callable that returns a new instance (like the type's initializer). This will allow us to instead return a factory function (which is not aType
) in future. For a given@component def C()
, this means:- You should not use
C
as a type anymore. For instance, replaceisinstance(foo, C)
with something else. Depending on your use case, if you just want to know whether it's a component, then useisinstance(foo, tfx.types.BaseComponent)
orisinstance(foo, tfx.types.BaseFunctionalComponent)
. If you want to know which component it is, check its.id
instead. Existing such checks will break type checking today and may additionally break at runtime in future, if we migrate to a factory function. - You can continue to use
C.test_call()
like before, and it will continue to be supported in future. - Any type declarations using
foo: C
break and must be replaced withfoo: tfx.types.BaseComponent
orfoo: tfx.types.BaseFunctionalComponent
. - Any references to static class members like
C.EXECUTOR_SPEC
breaks type checking today and should be migrated away from. In particular, for.EXECUTOR_SPEC.executor_class().Do()
in unit tests, use.test_call()
instead. - If your code previously asserted a wrong type declaration on
C
, this can now lead to (justified) type checking errors that were previously hidden due toC
being of typeAny
.
- You should not use
ph.to_list()
was renamed toph.make_list()
for consistency.
- Deprecated python 3.8
- Fixed a synchronization bug in google_cloud_ai_platform tuner.
- Print best tuning trials only from the chief worker of google_cloud_ai_platform tuner.
- Add a kpf dependency in the docker-image extra packages.
- Fix BigQueryExampleGen failure without custom_config.
Package Name | Version Constraints | Previously (in v1.14.0 ) |
Comments |
---|---|---|---|
keras-tuner |
>=1.0.4,<2 |
>=1.0.4,<2,!=1.4.0,!=1.4.1 |
|
packaging |
>=22 |
>=20,<21 |
|
attrs |
19.3.0,<24 |
19.3.0,<22 |
|
google-cloud-bigquery |
>=3,<4 |
>=2.26.0,<3 |
|
tensorflow |
>=2.13,<2.14 |
>=2.15,<2.16 |
|
tensorflow-decision-forests |
>=1.0.1,<2 |
>=1.0.1,<1.9 |
|
tensorflow-hub |
>=0.15.0,<0.16 |
>=0.9.0,<0.14 |
|
tensorflow-serving |
>=2.15,<2.16 |
>=1.15,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,!=2.7.*,!=2.8.*,<3 |
|
kfp-pipeline-spec |
>0.1.13,<0.2 |
>=0.1.10,<0.2 |
- Added python 3.10 support.
- Support
TypedDict
as a native output annotation for@component
.OutputDict
is still supported but it is recommended to useTypedDict
instead.
Placeholder
(and_PlaceholderOperator
) are no longerJsonable
.- Optimize MLMD register type to one call in most time instead of two calls.
- N/A
- Replace "tf_estimator" with "tfma_eval" as the identifier for tfma EvalSavedModel. "tf_estimator" is now serves as the identifier for the normal estimator model with any signature (by default 'serving').
- For
@component
return type annotation, it is recommended to use a python nativeTypedDict
instead.
- Apply latest TFX image vulnerability resolutions (base OS and software updates)
Package Name | Version Constraints | Previously (in v1.13.0 ) |
Comments |
---|---|---|---|
tensorflow-hub |
>=0.9.0,<0.14 |
>=0.9.0,<0.13 |
|
pyarrow |
>=10,<11 |
>=6,<7 |
|
apache-beam |
>=2.40,<3 |
>=2.47,<3 |
|
scikit-learn |
>=1.0,<2 |
>=0.23,<0.24 |
|
google-api-core |
<3 |
<1.33 |
|
google-cloud-aiplatform |
>=1.6.2,<2 |
>=1.6.2,<1.18 |
|
tflite-support |
>=0.4.3,<0.4.5 |
>=0.4.2,<0.4.3 |
|
pyyaml |
>=6,<7 |
>=3.12,<6 |
Issue with installation of PyYaml 5.4.1. (yaml/pyyaml#724) |
tensorflow |
>=2.13,<2.14 |
>=2.12,<2.13 |
|
tensorflowjs |
>=4.5,<5 |
>=3.6.0,<4 |
- N/A
- Supported setting the container image at a component level for Kubeflow V2 Dag Runner.
- Conditional can be used from
tfx.dsl.Cond
(Givenfrom tfx import v1 as tfx
). - Dummy channel for testing can be constructed by
tfx.testing.Channel(artifact_type)
. placeholder.Placeholder.placeholders_involved()
was replaced withplaceholder.Placeholder.traverse()
.placeholder.Predicate.dependent_channels()
was replaced withchannel_utils.get_dependent_channels(Placeholder)
.placeholder.Predicate.encode_with_keys(...)
was replaced withchannel_utils.encode_placeholder_with_channels(Placeholder, ...)
.placeholder.Predicate.from_comparison()
removed (was deprecated)- enable
external_pipeline_artifact_query
for querying artifact within one pipeline - Support
InputArtifact[List[Artifact]]
annotation in Python function custom component
- N/A
- Deprecate python 3.7 support
- Support to task type "workerpool1" of CLUSTER_SPEC in Vertex AI training's service according to the changes of task type in Tuner component.
- Propagates unexpected import failures in the public v1 module.
Package Name | Version Constraints | Previously (in v1.12.0 ) |
Comments |
---|---|---|---|
click |
>=7,<9 |
>=7,<8 |
|
ml-metadata |
~=1.13.1 |
~=1.12.0 |
Synced release train |
protobuf |
>=3.13,<4 |
>=3.20.3,<5 |
To support TF 2.12 |
struct2tensor |
~=0.44.0 |
~=0.43.0 |
Synced release train |
tensorflow |
~=2.12.0 |
>=1.15.5,<2 or ~=2.11.0 |
|
tensorflow-data-validation |
~=1.13.0 |
~=1.12.0 |
Synced release train |
tensorflow-model-analysis |
~=0.44.0 |
~=0.43.0 |
Synced release train |
tensorflow-transform |
~=1.13.0 |
~=1.12.0 |
Synced release train |
tfx-bsl |
~=1.13.0 |
~=1.12.0 |
Synced release train |
- Added page for TFX-Addons
- N/A
- N/A
- N/A
- N/A
- N/A
- ExampleValidator and DistributionValidator now support custom validations.
Package Name | Version Constraints | Previously (in v1.11.0 ) |
Comments |
---|---|---|---|
tensorflow |
~=2.11.0 |
>=1.15.5,<2 or ~=2.10.0 |
|
tensorflow-decision-forests |
>=1.0.1,<2 |
==1.0.1 |
Make it compatible with more TF versions. |
ml-metadata |
~=1.12.0 |
~=1.11.0 |
Synced release train |
struct2tensor |
~=0.43.0 |
~=0.42.0 |
Synced release train |
tensorflow-data-validation |
~=1.12.0 |
~=1.11.0 |
Synced release train |
tensorflow-model-analysis |
~=0.43.0 |
~=0.42.0 |
Synced release train |
tensorflow-transform |
~=1.12.0 |
~=1.11.0 |
Synced release train |
tfx-bsl |
~=1.12.0 |
~=1.11.0 |
Synced release train |
- N/A
-
This is the last version that supports TensorFlow 1.15.x. TF 1.15.x support will be removed in the next version. Please check the TF2 migration guide to migrate to TF2.
-
Artifact/Channel properties now support the new MLMD PROTO property type.
-
Supports environment variables in the placeholder expression. This placeholder can be used to generate beam_pipeline_args dynamically.
-
Update the TFMA Colab to utilize the DataFrame API to render metrics.
- Custom artifact types in kubeflow will encode
artifact.TYPE_NAME
as the schema title for the artifact instead of the class import path.
- N/A
- N/A
- N/A
- Moved
tflite-support
related dependencies from[examples]
to a separate[tflite-support]
extra. - Moved
flax
related dependencies from[examples]
to a separate[flax]
extra. - Statistics gen and Schema gen now crash on empty input examples and statistics respectively.
- Importer will now check that an existing artifact has the same type as the intended output before reusing the existing artifact.
- Importer will now use the most recently created artifact when reusing an existing artifact instead of the one with the highest ID.
- Proto placeholder now works with proto files that have non-trivial transitive dependencies.
- Adding tutorials for recommenders and ranking
Package Name | Version Constraints | Previously (in v1.10.0 ) |
Comments |
---|---|---|---|
tensorflow |
>=1.15.5,<2 or ~=2.10.0 |
>=1.15.5,<2 or ~=2.9.0 |
|
tflite-support |
~=0.4.2 |
>=0.1.0a1,<0.2.1 |
Update to a TF-2.10 compatible version. |
google-cloud-aiplatform |
>=1.6.2,<1.18 |
>=1.6.2,<2 |
Added to help pip dependency resolution. |
ml-metadata |
~=1.11.0 |
~=1.10.0 |
Synced release train |
struct2tensor |
~=0.42.0 |
~=0.41.0 |
Synced release train |
tensorflow-data-validation |
~=1.11.0 |
~=1.10.0 |
Synced release train |
tensorflow-model-analysis |
~=0.42.0 |
~=0.41.0 |
Synced release train |
tensorflow-transform |
~=1.11.0 |
~=1.10.0 |
Synced release train |
tfx-bsl |
~=1.11.0 |
~=1.10.0 |
Synced release train |
- N/A
- Saved tuner results in pandas
records
formatted JSON. - TFX Transform now supports
tf.SequenceExample
natively. The native path can be activated by providingTensorRepresentation
s in the Schema. - TFX Transform now supports reading raw and materializing transformed data in Apache Parquet format.
- ExampleDiff outputs statistics on the matching process, and optional counts of paired feature values.
- Allow lists and dicts to be passed into decorator components as parameters.
- N/A
- N/A
- N/A
- N/A
- Type hint on BaseComponent.inputs and BaseComponent.outputs corrected to be Channel subclasses.
- Added
input_optional
parameter toChannelParameter
. This allows component authors to declare that even if a channel isoptional
, if it is provided during pipeline definition time, then it must have resolved inputs during run time. - Allow latest
apache-airflow
2.x versions. - Output artifacts from multiple invocations of the same component are given unique names, avoiding duplication errors, especially in the InteractiveContext.
Package Name | Version Constraints | Previously (in v1.9.0 ) |
Comments |
---|---|---|---|
google-api-core |
<1.33 |
N/A | Added to help pip dependency resolution. google-api-core was already a transitive dependency. |
apache-beam[gcp] |
>=2.40,<3 |
>=2.38,<3 |
Synced release train |
attrs |
>=19.3.0,<22 |
>=19.3.0,<21 |
Allow more recent versions |
pyarrow |
>=6,<7 |
>=1,<6 |
Synced release train |
ml-metadata |
~=1.10.0 |
~=1.9.0 |
Synced release train |
struct2tensor |
~=0.41.0 |
~=0.40.0 |
Synced release train |
tensorflow-data-validation |
~=1.10.0 |
~=1.9.0 |
Synced release train |
tensorflow-model-analysis |
~=0.41.0 |
~=0.40.0 |
Synced release train |
tensorflow-transform |
~=1.10.1 |
~=1.9.0 |
Synced release train |
tfx-bsl |
~=1.10.1 |
~=1.9.0 |
Synced release train |
- N/A
- Added Json value artifact.
- Added example for using ExampleDiff.
- Allow lists and dicts to be consumed and produced by decorator components as input and output JsonValue artifacts.
- N/A
- N/A
- N/A
- N/A
- N/A
Package Name | Version Constraints | Previously (in v1.8.0 ) |
Comments |
---|---|---|---|
tensorflow |
>=1.15.5,<2 or ~=2.9.0 |
>=1.15.5,<2 or ~=2.8.0 |
|
tensorflow-ranking |
~=0.5.0 |
~=0.3.0 |
Required for TF 2.9 |
typing-extensions |
>=3.10.0.2,<5 |
N/A | For typing utilities |
ml-metadata |
~=1.9.0 |
~=1.8.0 |
Synced release train |
struct2tensor |
~=0.40.0 |
~=0.39.0 |
Synced release train |
tensorflow-data-validation |
~=1.9.0 |
~=1.8.0 |
Synced release train |
tensorflow-model-analysis |
~=0.40.0 |
~=0.39.0 |
Synced release train |
tensorflow-serving-api |
>=1.15,<3 or ~=2.9.0 |
>=1.15,<3 or ~=2.8.0 |
|
tensorflow-transform |
~=1.9.0 |
~=1.8.0 |
Synced release train |
tfx-bsl |
~=1.9.0 |
~=1.8.0 |
Synced release train |
- N/A
- Added experimental exit_handler support for KubeflowDagRunner.
- Enabled custom labels to be submitted to CAIP training jobs.
- Enabled custom resource-setting (vCPU and RAM) for containers orchestrating on Vertex AI.
- N/A
- N/A
- N/A
LatestBlessedModelStrategy
gracefully handles the case where there are no blessed model at all (e.g. first run).- Fix that the resolver with custom
ResolverStrategy
(assume correctly packaged) fails. - Fixed
ElwcBigQueryExampleGen
data serializiation error that was causing an assertion failure on Beam. - Added dark mode styling support for InteractiveContext notebook formatters.
- (Python 3.9+) Supports
list
anddict
in type definition of execution properties. - Populate Artifact proto
name
field when name is set on the Artifact python object. - Temporarily capped
apache-airflow
version to 2.2.x to avoid dependency conflict. We will rollback this change oncekfp
releases a new version. - Fixed a compatibility issue with apache-airflow 2.3.0 that is failing with "unexpected keyword argument 'default_args'".
- StatisticsGen will raise an error if unsupported StatsOptions (i.e., generators or experimental_slice_functions) are passed.
- Fixed a bug in the Artifact attribute setter that was causing the corresponding getter not to return a value for properties of type JSON_VALUE.
Package Name | Version Constraints | Previously (in v1.7.0 ) |
Comments |
---|---|---|---|
apache-beam[gcp] |
>=2.38,<3 |
>=2.36,<3 |
Synced release train |
- N/A
- Added support for list-type Placeholder.
- Added support for function-based custom component with beam pipeline.
- N/A
- N/A
- Removed the already-deprecated components.ImporterNode, should use v1.dsl.Importer instead.
- Deprecated Channel property setters. Use constructor argument instead.
- Fixed the cluster spec error in CAIP Tuner on Vertex when
num_parallel_trials = 1
- Replaced deprecated assertDictContainsSubset with assertLessEqual(itemsA, itemsB).
- Updating Keras tutorial to make better use of Keras, and better feature engineering.
- Merges KFP UI Metadata file if it already exists. Now components can produce their own UI results and it will be merged with existing visualization.
- Switch Transform component to always use sketch when computing top-k stats.
Package Name | Version Constraints | Previously (in v1.6.0 ) |
Comments |
---|---|---|---|
apache-beam[gcp] |
~=2.36 |
~=2.35 |
Synced release train |
google-cloud-aiplatform |
>=1.6.2,<2 |
>=1.5.0,<2 |
|
ml-metadata |
~=1.7.0 |
~=1.6.0 |
Synced release train |
struct2tensor |
~=0.38.0 |
~=0.37.0 |
Synced release train |
tensorflow |
>=1.15.5,<2 or ~=2.8.0 |
>=1.15.5,<2 or ~=2.7.0 |
|
tensorflow-data-validation |
~=1.7.0 |
~=1.6.0 |
Synced release train |
tensorflow-decision-forests |
==0.2.4 |
==0.2.1 |
|
tensorflow-model-analysis |
~=0.38.0 |
~=0.37.0 |
Synced release train |
tensorflow-serving-api |
>=1.15,<3 or ~=2.8.0 |
>=1.15,<3 or ~=2.7.0 |
|
tensorflow-transform |
~=1.7.0 |
~=1.6.0 |
Synced release train |
tfx-bsl |
~=1.7.0 |
~=1.6.0 |
Synced release train |
- N/A
- N/A
- N/A
- N/A
- N/A
- N/A
Package Name | Version Constraints | Previously (in v1.6.0 ) |
Comments |
---|---|---|---|
tensorflow |
>=1.15.5,<2 or ~=2.7.0 or ~=2.8.0 |
>=1.15.5,<2 or ~=2.7.0 |
- N/A
- N/A
- N/A
- N/A
- N/A
- Fixed
Pusher
issue that didn't copy files other thansaved_model.pb
.
- N/A
- Added experimental support for TensorFlow Decision Forests models.
- Added Boolean type value artifacts.
- Function components defined with
@component
may now have optional/nullable primitive type return values whenOptional[T]
is used in the return type OutputDict. - Supported endpoint overwrite for CAIP Tuner. Users can use the
keras-tuner
module or any tuner that implements thekeras_tuner.Tuner
interface for (parallel) tuning on Vertex.
- N/A
- N/A
- N/A
- Pusher now copies the
saved_model.pb
file at last to prevent loading SavedModel on invalid (partially available) directory state. - Always disable caching for exit handlers in Kubeflow V2 runner to reflect latest status of dependent dag.
Package Name | Version Constraints | Previously (in v1.5.0 ) |
Comments |
---|---|---|---|
tensorflow |
>=1.15.5,<2 or ~=2.7.0 |
>=1.15.2,<2 or ~=2.7.0 |
|
numpy |
~=1.16 |
>=1.16,<1.20 |
|
apache-beam[gcp] |
~=2.35 |
~=2.34 |
|
kfp |
~=1.8.5 |
>=1.6.1,<1.8.2,!=1.7.2 |
|
absl-py |
>=0.9,<2 |
>=0.9,<0.13 |
|
tfx-bsl |
~=1.6.0 |
~=1.5.0 |
Synced release train |
tensorflow-data-validation |
~=1.6.0 |
~=1.5.0 |
Synced release train |
tensorflow-transform |
~=1.6.0 |
~=1.5.0 |
Synced release train |
ml-metadata |
~=1.6.0 |
~=1.5.0 |
Synced release train |
tensorflow-model-analysis |
~=0.37.0 |
~=0.36.0 |
Synced release train |
struct2tensor |
~=0.37.0 |
~=0.36.0 |
Synced release train |
- N/A
- Added support for partial pipeline run. Users can now run a subset of nodes in a pipeline while reusing artifacts generated in previous pipeline runs. This is supported in LocalDagRunner and BeamDagRunner, and is exposed via the TfxRunner API.
- Add dependency of tensorflow-io to unblock using S3 storage.
- N/A
- N/A
- N/A
- Increased docker timeout to 5 minutes for image building in CLI.
- Fixed KeyError when multiple Examples artifacts were used in Transform without materialization.
- Fixed error where Vertex Endpoints of the same name is not deduped
- Depends on
apache-beam[gcp]>=2.34,<3
. - Depends on
tensorflow>=1.15.2,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,<2.8
. - Depends on
tensorflow-serving-api>=1.15,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,!=2.6.*,<3
. - Depends on
ml-metadata>=1.5.0,<1.6.0
. - Depends on
struct2tensor>=0.36.0,<0.37.0
. - Depends on
tensorflow-data-validation>=1.5.0,<1.6.0
. - Depends on
tensorflow-model-analysis>=0.36.0,<0.37.0
. - Depends on
tensorflow-transform>=1.5.0,<1.6.0
. - Depends on
tfx-bsl>=1.5.0,<1.6.0
.
- N/A
- Supported endpoint overwrite for CAIP BulkInferrer.
- Added support for outputting and encoding
tf.RaggedTensor
s in TFX Transform component. - Added conditional for TFX running on KFPv2 (Vertex).
- Supported component level beam pipeline args for Vertex (KFPV2DagRunner).
- Support exit handler for TFX running on KFPv2 (Vertex).
- Added RangeConfig for QueryBasedExampleGen to select date using query pattern.
- Added support for union of Channels as input to standard TFX components. Users can use channel.union() to combine multiple Channels and use as input to these compnents. Artfacts resolved from these channels are expected to have the same type, and passed to components in no particular order.
- Calling
TfxRunner.run(pipeline)
with the Pipeline IR proto will no longer be supported. Please switch toTfxRunner.run_with_ir(pipeline)
instead. If you are callingTfxRunner.run(pipeline)
with the Pipeline object, this change should not affect you.
- N/A
- N/A
- Deprecated python3.6 support.
- Depends on
google-cloud-aiplatform>=1.5.0,<2
. - Depends on
pyarrow>=1,<6
. - Fixed FileBasedExampleGen driver for Kubeflow v2 (Vertex). Driver can update exec_properties for its executor now, which enables {SPAN} feature.
- example_gen.utils.dict_to_example now accepts Numpy types
- Updated pytest to include v6.x
- Depends on
apache-beam[gcp]>=2.33,<3
. - Depends on
ml-metadata>=1.4.0,<1.5.0
. - Depends on
struct2tensor>=0.35.0,<0.36.0
. - Depends on
tensorflow-data-validation>=1.4.0,<1.5.0
. - Depends on
tensorflow-model-analysis>=0.35.0,<0.36.0
. - Depends on
tensorflow-transform>=1.4.0,<1.5.0
. - Depends on
tfx-bsl>=1.4.0,<1.5.0
.
- N/A
- N/A
- N/A
- N/A
- N/A
- Depends on
tensorflow>=1.15.2,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,<2.7
.
- N/A
- N/A
- N/A
- N/A
- N/A
- Fixed endless waiting for Vertex Trainer.
- N/A
- N/A
- N/A
- N/A
- N/A
- Fixed Vertex Pusher by passing enable_vertex flag for deploying model.
- N/A
- TFX CLI now supports runtime parameter on Kubeflow, Vertex, and Airflow. Use it with '--runtime_parameter=<parameter_name>=<parameter_value>' flag. In the case of multiple runtime parameters, format is as follows: '--runtime_parameter=<parameter_name>=<parameter_value> --runtime_parameter =<parameter_name>=<parameter_value>'
- Added Manual node in the experimental orchestrator.
- Placeholders support index access and JSON serialization for list type execution properties.
- Added
ImportSchemaGen
which is a dedicated component to import a pre-defined schema file. ImportSchemaGen will replaceImporter
with simpler syntax and less constraints. You have to pass the file path to the schema file instead of the parent directory unlikeImporter
. - Updated GCP Vertex Client to support EncryptionSpec and Cloud Labels.
- N/A
- N/A
- The import name of KerasTuner has been changed from
kerastuner
tokeras_tuner
. The import name ofkerastuner
is still supported. A warning will occur when import fromkerastuner
, but does not affect the usage. - Upcoming deprecation - TFX 1.3.0 will be the last release to support Python 3.6. Starting with TFX 1.4.0 Python 3.6 will no longer be supported.
- The default job name for Google Cloud AI Training jobs was changed from 'tfx_YYYYmmddHHMMSS' to 'tfx_YYYYmmddHHMMSS_xxxxxxxx', where 'xxxxxxxx' is a random 8 digit hexadecimal string.
- Fix component to raise error if its input required channel (specified from ComponentSpec) has no artifacts in it.
- Fixed an issue where ClientOptions with regional endpoint was incorrectly left out in Vertex AI pusher.
- CLI now hides passed flags from user python files in "--pipeline-path". This will prevent errors when user python file tries reading and parsing flags.
- Fixed missing type information marker file 'py.typed'.
- Fixed handling of artifacts with no PROPERTIES in scripts/run_component.py
- Fixed passing non-string execution properties and artifact properties in
scripts/run_component.py* Depends on
apache-beam[gcp]>=2.32,<3
. - Depends on
google-cloud-bigquery>=1.28.0,<3
. - Depends on
jinja2>=2.7.3,<4
, i.e. now supports Jinja 3.x. - Depends on
keras-tuner>=1.0.4,<2
. - Depends on
kfp>=1.6.1,!=1.7.2,<1.8.2
in [kfp] extra. - Depends on
kfp-pipeline-spec>=>=0.1.10,<0.2
. - Depends on
ml-metadata>=1.3.0,<1.4.0
. - Depends on
struct2tensor>=0.34.0,<0.35.0
. - Depends on
tensorflow>=1.15.2,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,<3
. - Depends on
tensorflow-data-validation>=1.3.0,<1.4.0
. - Depends on
tensorflow-model-analysis>=0.34.1,<0.35.0
. - Depends on
tensorflow-serving-api>=1.15,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,!=2.5.*,<3
. - Depends on
tensorflow-transform>=1.3.0,<1.4.0
. - Depends on
tfx-bsl>=1.3.0,<1.4.0
. - Depends on 'google-cloud-aiplatform>=0.5.0,<2'.
- N/A
- N/A
- N/A
- N/A
- N/A
- Added support for a custom metadata-ui-json filename in KubeflowDagRunner.
- Fixed missing type information marker file 'py.typed'.
- N/A
- Added RuntimeParam support for Trainer's custom_config.
- TFX Trainer and Pusher now support Vertex, which can be enabled with
ENABLE_VERTEX_KEY
key incustom_config
.
- N/A
- N/A
- N/A
- Fixed the issue that kfp_pod_name is not generated as an execution property for Kubeflow Pipelines.
- Fixed issue when InputValuePlaceholder is used as component parameter in container based component.
- Depends on
kubernetes>=10.0.1,<13
CsvToExample
now supports multi-line strings.tfx.benchmarks
package was removed from the Python TFX wheel. This package is used only for benchmarking and not useful for end users.- Fixed the issue for fairness_indicator_thresholds support of Evaluator.
- Depends on
apache-beam[gcp]>=2.31,<3
. - Depends on
kfp-pipeline-spec>=0.1.8,<0.2
. - Depends on
ml-metadata>=1.2.0,<1.3.0
. - Depends on
struct2tensor>=0.33.0,<0.34.0
. - Depends on
tensorflow-data-validation>=1.2.0,<1.3.0
. - Depends on
tensorflow-model-analysis>=0.33.0,<0.34.0
. - Depends on
tensorflow-transform>=1.2.0,<1.3.0
. - Depends on
tfx-bsl>=1.2.0,<1.3.0
.
- N/A
To maintain version consistency among TFX Family libraries we skipped the 1.1.x release for TFX library.
- Added tfx.v1 Public APIs, please refer to API doc for details.
- Transform component now computes pre-transform and post-transform statistics
and stores them in new, indvidual outputs ('pre_transform_schema',
'pre_transform_stats', 'post_transform_schema', 'post_transform_stats',
'post_transform_anomalies'). This can be disabled by setting
disable_statistics=True
in the Transform component. - BERT cola and mrpc examples now demonstrate how to calculate statistics for NLP features.
- TFX CLI now supports
Vertex Pipelines.
use it with
--engine=vertex
flag. - Telemetry: Only first-party tfx component's executor telemetry will be
collected. All other executors will be recorded as
third_party_executor
. For labels longer than 63, keep first 63 characters (instead of last 63 characters before). - Supports text type (use proto json string format) RuntimeParam for protos.
- Combined/moved taxi's runtime_parameter, kubeflow_local and kubleflow_gcp example pipelines into one penguin_pipeline_kubeflow example
- Transform component now supports passing
stats_options_updater_fn
directly as well as through the module file. - Placeholders support accessing artifact property and custom property.
- Removed the extra node information in IR for KubeflowDagRunner, to reduce size of generated IR.
- Removed unneccessary default values for required component input Channels.
- The
_PropertyDictWrapper
internal wrapper forcomponent.inputs
andcomponent.outputs
was removed:component.inputs
andcomponent.outputs
are now unwrapped dictionaries, and the attribute accessor syntax (e.g.components.outputs.output_name
) is no longer supported. Please use the dictionary indexing syntax (e.g.components.outputs['output_name']
) instead.
- N/A
- Apache Beam support is migrated from TFX Base Components and Executors to
dedicated Beam Components and Executors.
BaseExecutor
will no longer embedbeam_pipeline_args
. Custom executors for Beam powered components should now extend BaseBeamExecutor instead of BaseExecutor.
- Deprecated nested RuntimeParam for Proto, Please use text type (proto json string) RuntimeParam instead of Proto dict with nested RuntimeParam in it.
- Forces keyword arguments for AirflowComponent to make it compatible with Apache Airflow 2.1.0 and later.
- Fixed issue where passing
analyzer_cache
totfx.components.Transform
before there are any Transform cache artifacts published would fail. - Included type information according to PEP-561. However, protobuf generated
files don't have type information, and you might need to ignore errors from
them. For example, if you are using
mypy
, see the related doc. - Removed
six
dependency. - Depends on
apache-beam[gcp]>=2.29,<3
. - Depends on
google-cloud-bigquery>=1.28.0,<2.21
- Depends on
ml-metadata>=1.0.0,<1.1.0
. - Depends on
protobuf>=3.13,<4
. - Depends on
struct2tensor>=0.31.0,<0.32.0
. - Depends on
tensorflow>=1.15.2,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,<3
. - Depends on
tensorflow-data-validation>=1.0.0,<1.1.0
. - Depends on
tensorflow-hub>=0.9.0,<0.13
. - Depends on
tensorflowjs>=3.6.0,<4
. - Depends on
tensorflow-model-analysis>=0.31.0,<0.32.0
. - Depends on
tensorflow-serving-api>=1.15,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,!=2.4.*,<3
. - Depends on
tensorflow-transform>=1.0.0,<1.1.0
. - Depends on
tfx-bsl>=1.0.0,<1.1.0
.
- Update the Guide of TFX to adopt 1.0 API.
- TFT and TFDV component documentation now describes how to configure pre-transform and post-transform statistics, which can be used for validating text features.
- N/A
- N/A
- N/A
- N/A
- Update resolver query in TFX -> KFP IR compiler with vertex placeholder syntax.
- N/A
- TFX CLI now supports
Vertex Pipelines.
use it with
--engine=vertex
flag.
- N/A
- N/A
- N/A
- Fix resolver artifact filter in TFX -> KFP IR compiler with OP filter syntax.
- Forces keyword arguments for AirflowComponent to make it compatible with Apache Airflow 2.1.0 and later.
- N/A
- Upgraded TFX to KFP compiler to use KFP IR schema version 2.0.0.
- InfraValidator can now produce a SavedModel with warmup requests. This feature is
enabled by setting
RequestSpec.make_warmup = True
. The SavedModel will be stored in the InfraBlessing artifact (blessing
output of InfraValidator). - Pusher's
model
input is now optional, andinfra_blessing
can be used instead to push the SavedModel with warmup requests, produced by an InfraValidator. Note that InfraValidator does not always create a SavedModel, and the producer InfraValidator must be configured withRequestSpec.make_warmup = True
in order to be pushed by a Pusher. - Support is added for the JSON_VALUE artifact property type, allowing storage of JSON-compatible objects as artifact metadata.
- Support is added for the KFP v2 artifact metadata field when executing using the KFP v2 container entrypoint.
- InfraValidator for Kubernetes now can override Pod manifest to customize annotations and environment variables.
- Allow Beam pipeline args to be extended by specifying
beam_pipeline_args
per component. - Support string RuntimeParameters on Airflow.
- User code specified through the
module_file
argument for the Evaluator, Transform, Trainer and Tuner components is now packaged as a pip wheel for execution. For Evaluator and Transform, these wheel packages are now installed on remote Apache Beam workers.
- CLI usage with kubeflow changed significantly. You MUST use the new:
--build-image
to build a container image when updating a pipeline with kubeflow engine.--build-target-image
flag in CLI is changed to--build-image
without any container image argument. TFX will auto detect the image specified in the KubeflowDagRunnerConfig class instance. For example,tfx pipeline create --pipeline-path=runner.py --endpoint=xxx --build-image tfx pipeline update --pipeline-path=runner.py --endpoint=xxx --build-image
--package-path
and--skaffold_cmd
flags were deleted. The compiled path can be specified when creating a KubeflowDagRunner class instance. TFX CLI doesn't depend on skaffold any more and use Docker SDK directly.- Specify the container image for KubeflowDagRunner in the
KubeflowDagRunnerConfig directly instead of reading it from an environment
variable. CLI will not set
KUBEFLOW_TFX_IMAGE
environment variable any more. See example. - Default orchestration engine of CLI was changed to
local
orchestrator frombeam
orchestrator. You can still usebeam
orchestrator with--engine=beam
flag. - Trainer now uses GenericExecutor as default. To use the previous Estimator based Trainer, please set custom_executor_spec to trainer.executor.Executor.
- Changed the pattern spec supported for QueryBasedDriver:
- @span_begin_timestamp: Start of span interval, Timestamp in seconds.
- @span_end_timestamp: End of span interval, Timestamp in seconds.
- @span_yyyymmdd_utc: STRING with format, e.g., '20180114', corresponding to the span interval begin in UTC.
- Removed the already deprecated compile() method on Kubeflow V2 Dag Runner.
- Removed project_id argument from KubeflowV2DagRunnerConfig which is not used and meaningless if not used with GCP.
- Removed config from LocalDagRunner's constructor, and dropped pipeline proto support from LocalDagRunner's run function.
- Removed input parameter in ExampleGen constructor and external_input in dsl_utils, which were called as deprecated in TFX 0.23.
- Changed the storage type of
span
andversion
custom property in Examples artifact from string to int. ResolverStrategy.resolve_artifacts()
method signature has changed to takeml_metadata.MetadataStore
object as the first argument.- Artifacts param is deprecated/ignored in Channel constructor.
- Removed matching_channel_name from Channel's constructor.
- Deleted all usages of instance_name, which was deprecated in version 0.25.0. Please use .with_id() method of components.
- Removed output channel overwrite functionality from all official components.
- Transform will use the native TF2 implementation of tf.transform unless TF2
behaviors are explicitly disabled. The previous behaviour can still be
obtained by setting
force_tf_compat_v1=True
.
- N/A
- RuntimeParameter usage for
module_file
and user-defined function paths is marked experimental. LatestArtifactsResolver
,LatestBlessedModelResolver
,SpansResolver
are renamed toLatestArtifactStrategy
,LatestBlessedModelStrategy
,SpanRangeStrategy
respectively.
- GCP compute project in BigQuery Pusher executor can be specified.
- New extra dependencies for convenience.
- tfx[airflow] installs all Apache Airflow orchestrator dependencies.
- tfx[kfp] installs all Kubeflow Pipelines orchestrator dependencies.
- tfx[tf-ranking] installs packages for TensorFlow Ranking. NOTE: TensorFlow Ranking only compatible with TF >= 2.0.
- Depends on
google-cloud-bigquery>=1.28.0,<3
. (This was already installed as a transitive dependency from the first release of TFX.) - Depends on
google-cloud-aiplatform>=0.5.0,<0.8
. - Depends on
ml-metadata>=0.30.0,<0.31.0
. - Depends on
portpicker>=1.3.1,<2
. - Depends on
struct2tensor>=0.30.0,<0.31.0
. - Depends on
tensorflow-data-validation>=0.30.0,<0.31.0
. - Depends on
tensorflow-model-analysis>=0.30.0,<0.31.0
. - Depends on
tensorflow-transform>=0.30.0,<0.31.0
. - Depends on
tfx-bsl>=0.30.0,<0.31.0
.
- N/A
- Added a simple query based driver that supports Span spec and static_range.
- Added e2e rolling window example/test for Span Resolver.
- Performance improvement in Transform by avoiding excess encodings and decodings when it materializes transformed examples or generates statistics (both enabled by default).
- Added an accessor (
.data_view_decode_fn
) to the decoder function wrapped in the DataView in TrainerFnArgs.data_accessor
. - Expanded the penguin example pipeline with instructions for using JAX/Flax in addition to TensorFlow/Keras to write and train the model. The support for JAX/Flax in TFX is still experimental.
- Updated CloudTuner KFP e2e example running on Google Cloud Platform with distributed tuning and GPU distributed training for each trial.
- Starting in this version, following artifacts will be stored in new format,
but artifacts produced by older versions can be read in a backwards
compatible way:
- Change split sub-folder format to 'Split-<split_name>', this applies to all artifacts that contain splits. Old format '<split_name>' can still be loaded by TFX.
- Change Model artifact's sub-folder name to 'Format-TFMA' for eval model and 'Format-Serving' for serving model. Old Model artifact format ('eval_model_dir'/'serving_model_dir') can still be loaded by TFX.
- Change ExampleStatistics artifact payload to binary proto FeatureStats.pb file. Old payload format (tfrecord stats_tfrecord file) can still be loaded by TFX.
- Change ExampleAnomalies artifact payload to binary proto SchemaDiff.pb file. Old payload format (text proto anomalies.pbtxt file) is deprecated as TFX doesn't have downstream components that take ExampleAnomalies artifact.
- CLI requires Apache Airflow 1.10.14 or later. If you are using an older version of airflow, you can still copy runner definition to the DAG directory manually and run using airflow UIs.
- N/A
- Deprecated input/output compatibility aliases for Transform and StatisticsGen.
- The
tfx_version
custom property of output artifacts is now set by the default publisher to the TFX SDK version. - Depends on
absl-py>=0.9,<0.13
. - Depends on
kfp-pipeline-spec>=0.1.7,<0.2
. - Depends on
ml-metadata>=0.29.0,<0.30.0
. - Depends on
packaging>=20,<21
. - Depends on
struct2tensor>=0.29.0,<0.30.0
. - Depends on
tensorflow-data-validation>=0.29.0,<0.30.0
. - Depends on
tensorflow-model-analysis>=0.29.0,<0.30.0
. - Depends on
tensorflow-transform>=0.29.0,<0.30.0
. - Depends on
tfx-bsl>=0.29.0,<0.30.0
.
- Simplified Apache Spark and Flink example deployment scripts by using Beam's SparkRunner and FlinkRunner classes.
- Upgraded example Apache Flink deployment to Flink 1.12.1.
- Upgraded example Apache Spark deployment to Spark 2.4.7.
- Added the "TFX Python function component" notebook tutorial.
- Publically released TFX docker image in tensorflow/tfx will use GPU compatible based TensorFlow images from Deep Learning Containers. This allow these images to be used with GPU out of box.
- Added an example pipeline for a ranking model (using
tensorflow_ranking)
at
tfx/examples/ranking
. More documentation will be available in future releases. - Added a spans_resolver that can resolve spans based on range_config.
- Custom arg key in
google_cloud_ai_platform.tuner.executor
is renamed toai_platform_tuning_args
fromai_platform_training_args
, to better distinguish usage with Trainer.
- N/A
- Deprecated input/output compatibility aliases for Transform and SchemaGen.
- Change Bigquery ML Pusher to publish the model to the user specified project instead of the default project from run time context.
- Depends on
apache-beam[gcp]>=2.28,<3
. - Depends on
ml-metadata>=0.28.0,<0.29.0
. - Depends on
kfp-pipeline-spec>=0.1.6,<0.2
. - Depends on
struct2tensor>=0.28.0,<0.29.0
. - Depends on
tensorflow-data-validation>=0.28.0,<0.29.0
. - Depends on
tensorflow-model-analysis>=0.28.0,<0.29.0
. - Depends on
tensorflow-transform>=0.28.0,<0.29.0
. - Depends on
tfx-bsl>=0.28.1,<0.29.0
.
- Published a migration instruction for legacy custom launcher developers.
- Updated the
tfx.components.evaluator.Evaluator
component to support TFMA's "model-agnostic" evaluation. Themodel
channel is now optional when constructing the component, which is useful when theexamples
channel provides tf.Examples containing both the labels and pre-computed model predictions, i.e. "model-agnostic" evaluation. - Supports different types of quantizations on TFLite conversion using
TFLITE_REWRITER by setting
quantization_optimizations
,quantization_supported_types
andquantization_enable_full_integer
. Flag definitions can be found here: Post-traning quantization. - Added automatic population of
tfdv.StatsOptions.vocab_paths
when computing statistics within the Transform component.
enable_quantization
from TFLITE_REWRITER is removed and setting quantization_optimizations = [tf.lite.Optimize.DEFAULT] will perform the same type of quantization, dynamic range quantization. Users of the TFLITE_REWRITER who do not enable quantization should be uneffected.- Default value for
infer_feature_shape
for SchemaGen changed fromFalse
toTrue
, as indicated in previous release log. The inferred schema might change if you do not specifyinfer_feature_shape
. It might leads to changes of the type of input features in Transform and Trainer code.
- N/A
- Pipeline information is not be stored on the local filesystem anymore using
Kubeflow Pipelines orchestration with CLI. Instead, CLI will always use the
latest version of the pipeline in the Kubeflow Pipeline cluster. All
operations will be executed based on the information on the Kubeflow
Pipeline cluster. There might be some left files on
${HOME}/tfx/kubeflow
or${HOME}/kubeflow
but those will not be used any more. - The
tfx.components.common_nodes.importer_node.ImporterNode
class has been moved totfx.dsl.components.common.importer.Importer
, with its old module path kept as a deprecated alias, which will be removed in a future version. - The
tfx.components.common_nodes.resolver_node.ResolverNode
class has been moved totfx.dsl.components.common.resolver.Resolver
, with its old module path kept as a deprecated alias, which will be removed in a future version. - The
tfx.dsl.resolvers.BaseResolver
class has been moved totfx.dsl.components.common.resolver.ResolverStrategy
, with its old module path kept as a deprecated alias, which will be removed in a future version. - Deprecated input/output compatibility aliases for ExampleValidator, Evaluator, Trainer and Pusher.
- Add error condition checks to BulkInferrer's
output_example_spec
. Previously, when theoutput_example_spec
did not include the correct spec definitions, the BulkInferrer would fail silently and output examples without predictions. - InfraValidator supports using alternative TensorFlow Serving image in case
deployed environment cannot reach the public internet (nor the docker hub).
Such alternative image should behave the same as official
tensorflow/serving
image such as the same model volume path, serving port, etc. - Executor in
tfx.extensions.google_cloud_ai_platform.pusher.executor
supported regional endpoint and machine_type. - Starting from this version, proto files which are used to generate
component-level configs are included in the
tfx
package directly. - The
tfx.dsl.io.fileio.NotFoundError
exception unifies handling of not- found errors across different filesystem plugin backends. - Fixes the serialization of zero-valued default when using
RuntimeParameter
on Kubeflow. - Depends on
apache-beam[gcp]>=2.27,<3
. - Depends on
ml-metadata>=0.27.0,<0.28.0
. - Depends on
numpy>=1.16,<1.20
. - Depends on
pyarrow>=1,<3
. - Depends on
kfp-pipeline-spec>=0.1.5,<0.2
in test and image. - Depends on
tensorflow>=1.15.2,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,<3
. - Depends on
tensorflow-data-validation>=0.27.0,<0.28.0
. - Depends on
tensorflow-model-analysis>=0.27.0,<0.28.0
. - Depends on
tensorflow-serving-api>=1.15,!=2.0.*,!=2.1.*,!=2.2.*,!=2.3.*,<3
. - Depends on
tensorflow-transform>=0.27.0,<0.28.0
. - Depends on
tfx-bsl>=0.27.0,<0.28.0
.
- N/A
- This a bug fix only version.
- N/A
- N/A
- N/A
- N/A
- Depends on
apache-beam[gcp]>=2.25,!=2.26,<2.29
. - Depends on
tensorflow-data-validation>=0.26.1,<0.27
.
- N/A
- This a bug fix only version.
- N/A
- N/A
- N/A
- N/A
- Automatic autoreload of underlying modules a single
_ModuleFinder
registered per module.
- N/A
- This a bug fix only version
- N/A
- N/A
- N/A
- N/A
- The
tfx.version
attribute was restored.
- N/A
- Supported output examples artifact for BulkInferrer which can be used to link with downstream training.
- TFX Transform switched to a (notably) faster and more accurate
implementation of
tft.quantiles
analyzer. - Added native TF 2 implementation of Transform. The default
behavior will continue to use Tensorflow's compat.v1 APIs. This can be
overriden by passing
force_tf_compat_v1=False
and enabling TF 2 behaviors. The default behavior for TF 2 will be switched to the new native implementation in a future release. - Added support for passing a callable to set pre/post transform statistic generation options.
- In addition to the "tfx" pip package, a dependency-light distribution of the core pipeline authoring functionality of TFX is now available as the "ml-pipelines-sdk" pip package. This package does not include first-party TFX components. The "tfx" pip package is still the recommended installation path for TFX.
- Migrated LocalDagRunner to the new IR stack.
- Wheel package building for TFX has changed, and users need to follow the [new TFX package build instructions] (https://github.com/tensorflow/tfx/blob/master/package_build/README.md) to build wheels for TFX.
- Added BigQueryToElwcExampleGen to take a query as input and generate ExampleListWithContext (ELWC) examples.
- N/A
- TrainerFnArgs is deprecated by FnArgs.
- Deprecated DockerComponentConfig class: user should set a DockerPlatformConfig
proto in
platform_config
usingwith_platform_config()
API instead.
- Official TFX container image's entrypoint is changed so the image can be used as a custom worker for Dataflow.
- In the published TFX container image, wheel files are now used to install
TFX, and the TFX source code has been moved to
/tfx/src
. - Added a skeleton of CLI support for Kubeflow V2 runner, and implemented support for pipeline operations.
- Added an experimental template to use with Kubeflow V2 runner.
- Added sanitization of user-specified pipeline name in Kubeflow V2 runner.
- Migrated
deployment_config
in Kubeflow V2 runner fromAny
proto message toStruct
, to ensure compatibility across different copies of the proto libraries. - The
tfx.dsl.io.fileio
filesystem handler will delegate totensorflow.io.gfile
for any unknown filesystem schemes if TensorFlow is installed. - Skipped ephemeral package when the beam flag 'worker_harness_container_image' is set.
- The
tfx.dsl.io.makedirs
call now succeeds if the directory already exists. - Fixed the component entrypoint, so that it creates the parent directory for the output metadata file before trying to write the data.
- Depends on
apache-beam[gcp]>=2.25,!=2.26,<3
. - Depends on
keras-tuner>=1,<1.0.2
. - Depends on
kfp-pipeline-spec>=0.1.3,<0.2
. - Depends on
ml-metadata>=0.26.0,<0.27.0
. - Depends on
tensorflow>=1.15.2,!=2.0.*,!=2.1.*,!=2.2.*,!=2.4.*,<3
. - Depends on
tensorflow-data-validation>=0.26,<0.27
. - Depends on
tensorflow-model-analysis>=0.26,<0.27
. - Depends on
tensorflow-serving>=1.15,!=2.0.*,!=2.1.*,!=2.2.*,!=2.4.*,<3
. - Depends on
tensorflow-transform>=0.26,<0.27
. - Depends on
tfx-bsl>=0.26.1,<0.27
.
- N/A
-
Supported multiple artifacts for Transform's input example and output transformed example channels.
-
Added support for processing specific spans in file-based ExampleGen with range configuration.
-
Added ContainerExecutableSpec in portable IR to support container components portable orchestrator.
-
Added Placeholder utility library. Placeholder can be used to represent not-yet-available value at pipeline authoring time.
-
Added support for the
tfx.dsl.io.fileio
pluggable filesystem interface, with initial support for local files and the Tensorflow GFile filesystem implementation. -
SDK and example code now uses
tfx.dsl.io.fileio
instead oftf.io.gfile
when possible for filesystem I/O implementation portability. -
From this release TFX will also be hosting nightly packages on https://pypi-nightly.tensorflow.org. To install the nightly package use the following command:
pip install --extra-index-url https://pypi-nightly.tensorflow.org/simple tfx
Note: These nightly packages are unstable and breakages are likely to happen. The fix could often take a week or more depending on the complexity involved for the wheels to be available on the PyPI cloud service. You can always use the stable version of TFX available on PyPI by running the command
pip install tfx
-
Added CloudTuner KFP e2e example running on Google Cloud Platform with distributed tuning.
-
Migrated BigQueryExampleGen to the new
ReadFromBigQuery
on all runners. -
Introduced Kubeflow V2 DAG runner, which is based on Kubeflow IR spec. Same as
KubeflowDagRunner
it will compile the DSL pipeline into a payload but not trigger the execution locally. -
Added compile time check for schema mismatch in Kubeflow V2 runner.
-
Added 'penguin' example. Penguin example uses Palmer Penguins dataset and classify penguin species using four numeric features.
-
Iris e2e examples are replaced by penguin examples.
-
TFX BeamDagRunner is migrated to use the tech stack built on top of IR. While this is no-op to users, it is a major step towards supporting more flexible TFX DSL semetic. Please refer to the RFC of IR to learn more details.
-
Supports forward compatibility when evolving TFX artifact types, which allows jobs of old release and new release run with the same MLMD instance.
-
Graduated the portable/beam_dag_runner.py to beam/beam_dag_runner.py
- Moved the directory that CLI stores pipeline information from
${HOME}/${ORCHESTRATOR} to ${HOME}/tfx/${ORCHESTRATOR}. For example,
"
/kubeflow" was changed to "/tfx/kubeflow". This directory is used to store pipeline information including pipeline ids in the Kubeflow Pipelines cluster which are needed to create runs or update pipelines. These files will be moved automatically when it is first used and no separate action is needed. See https://github.com/tensorflow/tfx/blob/master/docs/guide/cli.md for the detail.
- N/A
- N/A
- Modules under
tfx.components.base
have been deprecated and moved totfx.dsl.components.base
in preparation for releasing a pipeline authoring package without explicit Tensorflow dependency. - Deprecated setting
instance_name
at pipeline node level. Instead, users are encouraged to setid
explicitly of any pipeline node through newly added APIs.
- Added the LocalDagRunner to allow local pipeline execution without using Apache Beam. This functionality is in development.
- Introduced dependency to
tensorflow-cloud
Python package, with intention to separate out Google Cloud Platform specific extensions. - Depends on
mmh>=2.2,<3
in container image for potential performance improvement for Beam based hashes. - New extra dependencies
[examples]
is required to use codes inside tfx/examples. - Fixed the run_component script.
- Stopped depending on
WTForms
. - Fixed an issue with Transform cache and beam 2.24-2.25 in an interactive notebook that caused it to fail.
- Scripts - run_component - Added a way to output artifact properties.
- Fixed an issue resulting in incorrect cache miss to ExampleGen when no
beam_pipeline_args
is provided. - Changed schema as an optional input channel of Trainer as schema can be accessed from TFT graph too.
- Fixed an issue during recording of a component's execution where
"missing or modified key in exec_properties" was raised from MLMD when
exec_properties
both omitted an existing property and added a new property. - Supported users to set
id
of pipeline nodes directly. - Added a new template, 'penguin' which is simple subset of penguin examples, and uses the same Palmer Penguins dataset. The new template focused on easy ingestion of user's own data.
- Changed default data path for the taxi template from
tfx-template/data
totfx-template/data/taxi
. - Fixed a bug which crashes the pusher when infra validation did not pass.
- Depends on
apache-beam[gcp]>=2.25,<3
. - Depends on
attrs>=19.3.0,<21
. - Depends on
kfp-pipeline-spec>=0.1.2,<0.2
. - Depends on
kfp>=1.1.0,<2
. - Depends on
ml-metadata>=0.25,<0.26
. - Depends on
tensorflow-cloud>=0.1,<0.2
. - Depends on
tensorflow-data-validation>=0.25,<0.26
. - Depends on
tensorflow-hub>=0.9.0,<0.10
. - Depends on
tensorflow-model-analysis>=0.25,<0.26
. - Depends on
tensorflow-transform>=0.25,<0.26
. - Depends on
tfx-bsl>=0.25,<0.26
.
- N/A
- N/A
- Fixes issues where custom property access of a missing property created an invalid MLMD Artifact protobuf message.
- N/A
- N/A
- N/A
- N/A
- Use TFXIO and batched extractors by default in Evaluator.
- Supported custom split configuration for ExampleGen and its downstream components. Instead of hardcoded 'train' and 'eval' splits, TFX components now can process the custom splits generated by ExampleGen. For details, please refer to ExampleGen doc
- Added python 3.8 support.
- Supported CAIP Runtime 2.2 for online prediction pusher.
- Used 'python -m ' style for container entrypoints.
- Stopped depending on
google-resumable-media
. - Stopped depending on
Werkzeug
. - Depends on
absl-py>=0.9,<0.11
. - Depends on
apache-beam[gcp]>=2.24,<3
. - Depends on
ml-metadata>=0.24,<0.25
. - Depends on
protobuf>=3.12.2,<4
. - Depends on
tensorflow-data-validation>=0.24.1,<0.25
. - Depends on
tensorflow-model-analysis>=0.24.3,<0.25
. - Depends on
tensorflow-transform>=0.24.1,<0.25
. - Depends on
tfx-bsl>=0.24.1,<0.25
.
- N/A
- N/A
- N/A
- Deprecated python 3.5 support.
- This is a bug fix version (to resolve impossible dependency conflicts).
- N/A
- Stopped depending on
google-resumable-media
. - Depends on
apache-beam[gcp]>=2.24,<3
. - Depends on
tensorflow-data-validation>=0.23.1,<0.24
.
- N/A
- N/A
- N/A
- Deprecated Python 3.5 support.
- Added TFX DSL IR compiler that encodes a TFX pipeline into a DSL proto.
- Supported feature based split partition in ExampleGen.
- Added the ConcatPlaceholder to tfx.dsl.component.experimental.placeholders.
- Changed Span information as a property of ExampleGen's output artifact. Deprecated ExampleGen input (external) artifact.
- Added ModelRun artifact for Trainer for storing training related files,
e.g., Tensorboard logs. Trainer's Model artifact now only contain pure
models (check
tfx/utils/path_utils.py
for details). - Added support for
tf.train.SequenceExample
in ExampleGen:- ImportExampleGen now supports
tf.train.SequenceExample
importing. - base_example_gen_executor now supports
tf.train.SequenceExample
as output payload format, which can be utilized by custom ExampleGen.
- ImportExampleGen now supports
- Added Tuner component and its integration with Google Cloud Platform as the execution and hyperparemeter optimization backend.
- Switched Transform component to use the new TFXIO code path. Users may potentially notice large performance improvement.
- Added support for primitive artifacts to InputValuePlaceholder.
- Supported multiple artifacts for Trainer and Tuner's input example Channel.
- Supported split configuration for Trainer and Tuner.
- Supported split configuration for Evaluator.
- Supported split configuration for StatisticsGen, SchemaGen and
ExampleValidator. SchemaGen will now use all splits to generate schema
instead of just using
train
split. ExampleValidator will now validate all splits against given schema instead of just validatingeval
split. - Component authors now can create a TFXIO instance to get access to the
data through
tfx.components.util.tfxio_utils
. As TFX is going to support more data payload formats and data container formats, usingtfxio_utils
is encouraged to avoid dealing directly with each combination. TFXIO is the interface of Standardized TFX Inputs. - Added experimental BaseStubExecutor and StubComponentLauncher to test TFX pipelines.
- Added experimental TFX Pipeline Recorder to record output artifacts of the pipeline.
- Supported multiple artifacts in an output Channel to match a certain input Channel's artifact count. This enables Transform component to process multiple artifacts.
- Transform component's transformed examples output is now optional (enabled
by default). This can be disabled by specifying parameter
materialize=False
when constructing the component. - Supported
Version
spec in input config for file based ExampleGen. - Added custom config to Transform component and made it available to pre-processing fn.
- Supported custom extractors in Evaluator.
- Deprecated tensorflow dependency from MLMD python client.
- Supported
Date
spec in input config for file based ExampleGen. - Enabled analyzer cache optimization in the Transform component:
- specify
analyzer_cache
to use the cache generated from a previous run. - specify parameter
disable_analyzer_cache=True
(False by default) to disable cache (won't generate cache output).
- specify
- Added support for width modifiers in Span and Version specs for file based ExampleGen.
- Added Tuner component to Iris e2e example.
- Relaxed the rule that output artifact uris must be newly created. This is a temporary workaround to make retry work. We will introduce a more comprehensive solution for idempotent execution.
- Made evaluator output optional (while still recommended) for pusher.
- Moved BigQueryExampleGen to
tfx.extensions.google_cloud_big_query
. - Moved BigQuery ML Pusher to
tfx.extensions.google_cloud_big_query.pusher
. - Removed Tuner from custom_components/ as it's supported under components/ now.
- Added support of non tf.train.Example protos as internal data payload format by ImportExampleGen.
- Used thread local storage for
label_utils.scoped_labels()
to make it thread safe. - Requires Bazel to build TFX source code.
- Upgraded python version in TFX docker images to 3.7. Older version of
python (2.7/3.5/3.6) is not available anymore in
tensorflow/tfx
images on docker hub. Virtualenv is not used anymore. - Stopped requiring
avro-python3
. - Depends on
absl-py>=0.7,<0.9
. - Depends on
apache-beam[gcp]>=2.23,<3
. - Depends on
pyarrow>=0.17,<0.18
. - Depends on
attrs>=19.3.0,<20
. - Depends on
ml-metadata>=0.23,<0.24
. - Depends on
tensorflow>=1.15.2,!=2.0.*,!=2.1.*,!=2.2.*,<3
.- Note: Dependency like
tensorflow-transform
might impose a narrower range oftensorflow
.
- Note: Dependency like
- Depends on
tensorflow-data-validation>=0.23,<0.24
. - Depends on
tensorflow-model-analysis>=0.23,<0.24
. - Depends on
tensorflow-serving-api>=1.15,!=2.0.*,!=2.1.*,!=2.2.*,<3
. - Depends on
tensorflow-transform>=0.23,<0.24
. - Depends on
tfx-bsl>=0.23,<0.24
. - Added execution_result_pb2.ExecutorOutput as an Optional return value of BaseExecutor. This change is backward compatible to all existing executors.
- Added executor_output_uri and stateful_working_dir to Executor's context.
- Changed the URIs of the value artifacts to point to files.
- De-duplicated the tfx.dsl.component.experimental.executor_specs.CommandLineArgumentType union type in favor of tfx.dsl.component.experimental.placeholders.CommandLineArgumentType
- Moved BigQueryExampleGen to
tfx.extensions.google_cloud_big_query
. The previous module path fromtfx.components
is not available anymore. This is a breaking change. - Moved BigQuery ML Pusher to
tfx.extensions.google_cloud_big_query.pusher
. The previous module path fromtfx.extensions.google_cloud_big_query_ml
is not available anymore. - Updated beam pipeline args, users now need to set both
direct_running_mode
anddirect_num_workers
explicitly for multi-processing. - Added required 'output_data_format' execution property to FileBaseExampleGen.
- Changed ExampleGen to take a string as input source directly instead of a
Channel of external artifact:
- Previously deprecated
input_base
Channel is changed to string type instead of Channel. This is a breaking change, users should pass string directly toinput_base
.
- Previously deprecated
- Fully removed csv_input and tfrecord_input in dsl_utils. This is a breaking
change, users should pass string directly to
input_base
.
- Changed GetInputSourceToExamplePTransform interface by removing input_dict. This is a breaking change, custom ExampleGens need to follow the interface change.
- Changed ExampleGen to take a string as input source directly instead of a
Channel of external artifact:
input
Channel is deprecated. The use ofinput
is valid but should change to string typeinput_base
ASAP.
- N/A
- ExternalArtifact and
external_input
function are deprecated. The use ofexternal_input
with ExampleGeninput
is still valid but should change to useinput_base
ASAP. - Note: We plan to remove Python 3.5 support after this release.
- N/A
- Reuse Examples artifact type introduced in TFX 0.23 to allow older release jobs running together with TFX 0.23+ release.
- N/A
- N/A
- N/A
- N/A
- N/A
- Depends on 'tensorflowjs>=2.0.1.post1,<3' for
[all]
dependency. - Fixed the name of the usage telemetry when tfx templates are used.
- Depends on
tensorflow-data-validation>=0.22.2,<0.23.0
. - Depends on
tensorflow-model-analysis>=0.22.2,<0.23.0
. - Depends on
tfx-bsl>=0.22.1,<0.23.0
. - Depends on
ml-metadata>=0.22.1,<0.23.0
.
- N/A
- N/A
- N/A
- N/A
- Introduced experimental Python function component decorator (
@component
decorator undertfx.dsl.component.experimental.decorators
) allowing Python function-based component definition. - Added the experimental TemplatedExecutorContainerSpec executor class that supports structural placeholders (not Jinja placeholders).
- Added the experimental function "create_container_component" that simplifies creating container-based components.
- Implemented a TFJS rewriter.
- Added the scripts/run_component.py script which makes it easy to run the component code and executor code. (Similar to scripts/run_executor.py)
- Added support for container component execution to BeamDagRunner.
- Introduced experimental generic Artifact types for ML workflows.
- Added support for
float
execution properties.
- Migrated BigQueryExampleGen to the new (experimental)
ReadFromBigQuery
PTramsform when not using Dataflow runner. - Enhanced add_downstream_node / add_upstream_node to apply symmetric changes when being called. This method enables task-based dependencies by enforcing execution order for synchronous pipelines on supported platforms. Currently, the supported platforms are Airflow, Beam, and Kubeflow Pipelines. Note that this API call should be considered experimental, and may not work with asynchronous pipelines, sub-pipelines and pipelines with conditional nodes.
- Added the container-based sample pipeline (download, filter, print)
- Removed the incomplete cifar10 example.
- Removed
python-snappy
from[all]
extra dependency list. - Tests depends on
apache-airflow>=1.10.10,<2
; - Removed test dependency to tzlocal.
- Fixes unintentional overriding of user-specified setup.py file for Dataflow jobs when running on KFP container.
- Made ComponentSpec().inputs and .outputs behave more like real dictionaries.
- Depends on
kerastuner>=1,<2
. - Depends on
pyyaml>=3.12,<6
. - Depends on
apache-beam[gcp]>=2.21,<3
. - Depends on
grpcio>=2.18.1,<3
. - Depends on
kubernetes>=10.0.1,<12
. - Depends on
tensorflow>=1.15,!=2.0.*,<3
. - Depends on
tensorflow-data-validation>=0.22.0,<0.23.0
. - Depends on
tensorflow-model-analysis>=0.22.1,<0.23.0
. - Depends on
tensorflow-transform>=0.22.0,<0.23.0
. - Depends on
tfx-bsl>=0.22.0,<0.23.0
. - Depends on
ml-metadata>=0.22.0,<0.23.0
. - Depends on 'tensorflowjs>=2.0.1.post1,<3' for
[all]
dependency. - Fixed a bug in
io_utils.copy_dir
which prevent it to work correctly for nested sub-directories.
- Changed custom config for the Do function of Trainer and Pusher to accept
a JSON-serialized dict instead of a dict object. This also impacts all the
Do functions under
tfx.extensions.google_cloud_ai_platform
andtfx.extensions.google_cloud_big_query_ml
. Note that this breaking change occurs at the signature of the executor's Do function. Therefore, if the user did not customize the Do function, and the compile time SDK version is aligned with the run time SDK version, previous pipelines should still work as intended. If the user is using a custom component with customized Do function,custom_config
should be assumed to be a JSON-serialized string from next release. - For users of BigQueryExampleGen,
--temp_location
is now a required Beam argument, even for DirectRunner. Previously this argument was only required for DataflowRunner. Note that the specified value of--temp_location
should point to a Google Cloud Storage bucket. - Revert current per-component cache API (with
enable_cache
, which was only available in tfx>=0.21.3,<0.22), in preparing for a future redesign.
- Converted the BaseNode class attributes to the constructor parameters. This won't affect any components derived from BaseComponent.
- Changed the encoding of the Integer and Float artifacts to be more portable.
- Added concept guides for understanding TFX pipelines and components.
- Added guides to building Python function-based components and container-based components.
- Added BulkInferrer component and TFX CLI documentation to the table of contents.
- Deprecating Py2 support
- N/A
- Reuse Examples artifact type introduced in TFX 0.23 to allow older release jobs running together with TFX 0.23+ release.
- Removed python-snappy from [all] extra dependency list.
- N/A
- N/A
- N/A
- N/A
- N/A
- Fixed InfraValidator signal handling bug on BeamDagRunner.
- Dropped "Type" suffix from primitive type artifact names (Integer, Float, String, Bytes).
- N/A
- N/A
- N/A
- N/A
- Added run/pipeline link when creating runs/pipelines on KFP through TFX CLI.
- Added support for
ValueArtifact
, whose attributevalue
allows users to access the content of the underlying file directly in the executor. Support Bytes/Integer/String/Float type. Note: interactive resolution does not support this for now. - Added InfraValidator component that is used as an early warning layer before pushing a model into production.
- Starting this version, TFX will only release python3 packages.
- Replaced relative import with absolute import in generated templates.
- Added a native keras model in the taxi template and the template now uses generic Trainer.
- Added support of TF 2.1 runtime configuration for AI Platform Prediction Pusher.
- Added support for using ML Metadata ArtifactType messages as Artifact classes.
- Changed CLI behavior to create new versions of pipelines instead of delete and create new ones when pipelines are updated for KFP. (Requires kfp >= 0.3.0)
- Added ability to enable quantization in tflite rewriter.
- Added k8s pod labels when the pipeline is executed via KubeflowDagRunner for better usage telemetry.
- Parameterized the GCP taxi pipeline sample for easily ramping up to full taxi dataset.
- Added support for hyphens(dash) in addition to underscores in CLI flags. Underscores will be supported as well.
- Fixed ill-formed underscore in the markdown visualization when running on KFP.
- Enabled per-component control for caching with enable_cache argument in each component.
- N/A
- N/A
- N/A
- N/A
- Updated
StatisticsGen
to optionally consume a schemaArtifact
. - Added support for configuring the
StatisticsGen
component via serializable parts ofStatsOptions
. - Added Keras guide doc.
- Changed Iris model_to_estimator e2e example to use generic Trainer.
- Demonstrated how TFLite is supported in TFX by extending MNIST example pipeline to also train a TFLite model.
- Fix the behavior of Trainer Tensorboard visualization when caching is used.
- Added component documentation and guide on using TFLite in TFX.
- Relaxed the PyYaml dependency.
- Model Validator (its functionality is now provided by the Evaluator).
- N/A
- N/A
- N/A
- Pipelines compiled using KubeflowDagRunner now defaults to using the gRPC-based MLMD server deployed in Kubeflow Pipelines clusters when performing operations on pipeline metadata.
- Added tfx model rewriting and tflite rewriter.
- Added LatestBlessedModelResolver as an experimental feature which gets the latest model that was blessed by model validator.
- The specific
Artifact
subclass that was serialized (if defined in the deserializing environment) will be used when deserializingArtifact
s and when readingArtifact
s from ML Metadata (previously, objects of the generictfx.types.artifact.Artifact
class were created in some cases). - Updated Evaluator's executor to support model validation.
- Introduced awareness of chief worker to Trainer's executor, in case running in distributed training cluster.
- Added a Chicago Taxi example with native Keras.
- Updated TFLite converter to work with TF2.
- Enabled filtering by artifact producer and output key in ResolverNode.
- Added --skaffold_cmd flag when updating a pipeline for kubeflow in CLI.
- Changed python_version to 3.7 when using TF 1.15 and later for Cloud AI Platform Prediction.
- Added 'tfx_runner' label for CAIP, BQML and Dataflow jobs submitted from TFX components.
- Fixed the Taxi Colab notebook.
- Adopted the generic trainer executor when using CAIP Training.
- Depends on 'tensorflow-data-validation>=0.21.4,<0.22'.
- Depends on 'tensorflow-model-analysis>=0.21.4,<0.22'.
- Depends on 'tensorflow-transform>=0.21.2,<0.22'.
- Fixed misleading logs in Taxi pipeline portable Beam example.
- N/A
- Remove "NOT_BLESSED" artifact.
- Change constants ARTIFACT_PROPERTY_BLESSED_MODEL_* to ARTIFACT_PROPERTY_BASELINE_MODEL_*.
- N/A
- N/A
- N/A
- TFX version 0.21.0 will be the last version of TFX supporting Python 2.
- Added experimental cli option
template
, which can be used to scaffold a new pipeline from TFX templates. Currently thetaxi
template is provided and more templates would be added in future versions. - Added support for
RuntimeParameter
s to allow users can specify templated values at runtime. This is currently only supported in Kubeflow Pipelines. Currently, only attributes inComponentSpec.PARAMETERS
and the URI of external artifacts can be parameterized (component inputs / outputs can not yet be parameterized). Seetfx/examples/chicago_taxi_pipeline/taxi_pipeline_runtime_parameter.py
for example usage. - Users can access the parameterized pipeline root when defining the
pipeline by using the
pipeline.ROOT_PARAMETER
placeholder in KubeflowDagRunner. - Users can pass appropriately encoded Python
dict
objects to specify protobuf parameters inComponentSpec.PARAMETERS
; these will be decoded into the proper protobuf type. Users can avoid manually constructing complex nested protobuf messages in the component interface. - Added support in Trainer for using other model artifacts. This enables scenarios such as warm-starting.
- Updated trainer executor to pass through custom config to the user module.
- Artifact type-specific properties can be defined through overriding the
PROPERTIES
dictionary of atypes.artifact.Artifact
subclass. - Added new example of chicago_taxi_pipeline on Google Cloud Bigquery ML.
- Added support for multi-core processing in the Flink and Spark Chicago Taxi PortableRunner example.
- Added a metadata adapter in Kubeflow to support logging the Argo pod ID as an execution property.
- Added a prototype Tuner component and an end-to-end iris example.
- Created new generic trainer executor for non estimator based model, e.g., native Keras.
- Updated to support passing
tfma.EvalConfig
in evaluator when calling TFMA. - Added an iris example with native Keras.
- Added an MNIST example with native Keras.
- Switched the default behavior of KubeflowDagRunner to not mounting GCP secret.
- Fixed "invalid spec: spec.arguments.parameters[6].name 'pipeline-root' is
not unique" error when the user include
pipeline.ROOT_PARAMETER
and run pipeline on KFP. - Added support for an hparams artifact as an input to Trainer in preparation for tuner support.
- Refactored common dependencies in the TFX dockerfile to a base image to improve the reliability of image building process.
- Fixes missing Tensorboard link in KubeflowDagRunner.
- Depends on
apache-beam[gcp]>=2.17,<2.18
- Depends on
ml-metadata>=0.21,<0.22
. - Depends on
tensorflow-data-validation>=0.21,<0.22
. - Depends on
tensorflow-model-analysis>=0.21,<0.22
. - Depends on
tensorflow-transform>=0.21,<0.22
. - Depends on
tfx-bsl>=0.21,<0.22
. - Depends on
pyarrow>=0.14,<0.15
. - Removed
tf.compat.v1
usage for iris and cifar10 examples. - CSVExampleGen: started using the CSV decoding utilities in
tfx-bsl
(tfx-bsl>=0.15.2
) - Fixed problems with Airflow tutorial notebooks.
- Added performance improvements for the Transform Component (for statistics generation).
- Raised exceptions when container building fails.
- Enhanced custom slack component by adding a kubeflow example.
- Allowed windows style paths in Transform component cache.
- Fixed bug in CLI (--engine=kubeflow) which uses hard coded obsolete image (TFX 0.14.0) as the base image.
- Fixed bug in CLI (--engine=kubeflow) which could not handle skaffold response when an already built image is reused.
- Allowed users to specify the region to use when serving with AI Platform.
- Allowed users to give deterministic job id to AI Platform Training job.
- System-managed artifact properties ("name", "state", "pipeline_name" and "producer_component") are now stored as ML Metadata artifact custom properties.
- Fixed loading trainer and transformation functions from python module files without the .py extension.
- Fixed some ill-formed visualization when running on KFP.
- Removed system info from artifact properties and use channels to hold info for generating MLMD queries.
- Rely on MLMD context for inter-component artifact resolution and execution publishing.
- Added pipeline level context and component run level context.
- Included test data for examples/chicago_taxi_pipeline in package.
- Changed
BaseComponentLauncher
to require the user to pass in an ML Metadata connection object instead of a ML Metadata connection config. - Capped version of Tensorflow runtime used in Google Cloud integration to 1.15.
- Updated Chicago Taxi example dependencies to Beam 2.17.0, Flink 1.9.1, Spark 2.4.4.
- Fixed an issue where
build_ephemeral_package()
used an incorrect path to locate thetfx
directory. - The ImporterNode now allows specification of general artifact properties.
- Added 'tfx_executor', 'tfx_version' and 'tfx_py_version' labels for CAIP, BQML and Dataflow jobs submitted from TFX components.
- Use '_' instead of '/' in feature names of several examples to avoid potential clash with namescope separator.
- N/A
- Standard artifact TYPE_NAME strings were reconciled to match their class
names in
types.standard_artifacts
. - The "split" property on multiple artifacts has been replaced with the JSON-encoded "split_names" property on a single grouped artifact.
- The execution caching mechanism was changed to rely on ML Metadata pipeline context. Existing cached executions will not be reused when running on this version of TFX for the first time.
- The "split" property on multiple artifacts has been replaced with the JSON-encoded "split_names" property on a single grouped artifact.
- Artifact type name strings to the
types.artifact.Artifact
andtypes.channel.Channel
classes are no longer supported; usage here should be replaced with references to the artifact subclasses defined intypes.standard_artfacts.*
or to custom subclasses oftypes.artifact.Artifact
.
- N/A
- Offered unified CLI for tfx pipeline actions on various orchestrators including Apache Airflow, Apache Beam and Kubeflow.
- Polished experimental interactive notebook execution and visualizations so they are ready for use.
- Added BulkInferrer component to TFX pipeline, and corresponding offline inference taxi pipeline.
- Introduced ImporterNode as a special TFX node to register external resource
into MLMD so that downstream nodes can use as input artifacts. An example
taxi_pipeline_importer.py
enabled by ImporterNode was added to showcase the user journey of user-provided schema (issue #571). - Added experimental support for TFMA fairness indicator thresholds.
- Demonstrated DirectRunner multi-core processing in Chicago Taxi example, including Airflow and Beam.
- Introduced
PipelineConfig
andBaseComponentConfig
to control the platform specific settings for pipelines and components. - Added a custom Executor of Pusher to push model to BigQuery ML for serving.
- Added KubernetesComponentLauncher to support launch ExecutorContainerSpec in a Kubernetes cluster.
- Made model validator executor forward compatible with TFMA change.
- Added Iris flowers classification example.
- Added support for serialization and deserialization of components.
- Made component launcher extensible to support launching components on multiple platforms.
- Simplified component package names.
- Introduced BaseNode as the base class of any node in a TFX pipeline DAG.
- Added docker component launcher to launch container component.
- Added support for specifying pipeline root in runtime when run on KubeflowDagRunner. A default value can be provided when constructing the TFX pipeline.
- Added basic span support in ExampleGen to ingest file based data sources that can be updated regularly by upstream.
- Branched serving examples under chicago_taxi_pipeline/ from chicago_taxi/ example.
- Supported beam arg 'direct_num_workers' for multi-processing on local.
- Improved naming of standard component inputs and outputs.
- Improved visualization functionality in the experimental TFX notebook interface.
- Allowed users to specify output file format when compiling TFX pipelines using KubeflowDagRunner.
- Introduced ResolverNode as a special TFX node to resolve input artifacts for downstream nodes. ResolverNode is a convenient way to wrap TFX Resolver, a logical unit for resolving input artifacts.
- Added cifar-10 example to demonstrate image classification.
- Added container builder feature in the CLI tool for container-based custom python components. This is specifically for the Kubeflow orchestration engine, which requires containers built with the custom python code.
- Demonstrated DirectRunner multi-core processing in Chicago Taxi example, including Airflow and Beam.
- Added Kubeflow artifact visualization of inputs, outputs and execution properties for components using a Markdown file. Added Tensorboard to Trainer components as well.
- Bumped test dependency to kfp (Kubeflow Pipelines SDK) to be at version 0.1.31.2.
- Fixed trainer executor to correctly make
transform_output
optional. - Updated Chicago Taxi example dependency tensorflow to version >=1.14.0.
- Updated Chicago Taxi example dependencies tensorflow-data-validation, tensorflow-metadata, tensorflow-model-analysis, tensorflow-serving-api, and tensorflow-transform to version >=0.14.
- Updated Chicago Taxi example dependencies to Beam 2.14.0, Flink 1.8.1, Spark 2.4.3.
- Adopted new recommended way to access component inputs/outputs as
component.outputs['output_name']
(previously, the syntax wascomponent.outputs.output_name
). - Updated Iris example to skip transform and use Keras model.
- Fixed the check for input artifact existence in base driver.
- Fixed bug in AI Platform Pusher that prevents pushes after first model, and not being marked as default.
- Replaced all usage of deprecated
tensorflow.logging
withabsl.logging
. - Used special user agent for all HTTP requests through googleapiclient and apitools.
- Transform component updated to use
tf.compat.v1
according to the TF 2.0 upgrading procedure. - TFX updated to use
tf.compat.v1
according to the TF 2.0 upgrading procedure. - Added Kubeflow local example pipeline that executes components in-cluster.
- Fixed a bug that prevents updating execution type.
- Fixed a bug in model validator driver that reads across pipeline boundaries when resolving latest blessed model.
- Depended on
apache-beam[gcp]>=2.16,<3
- Depended on
ml-metadata>=0.15,<0.16
- Depended on
tensorflow>=1.15,<3
- Depended on
tensorflow-data-validation>=0.15,<0.16
- Depended on
tensorflow-model-analysis>=0.15.2,<0.16
- Depended on
tensorflow-transform>=0.15,<0.16
- Depended on 'tfx_bsl>=0.15.1,<0.16'
- Made launcher return execution information, containing populated inputs, outputs, and execution id.
- Updated the default configuration for accessing MLMD from pipelines running in Kubeflow.
- Updated Airflow developer tutorial
- CSVExampleGen: started using the CSV decoding utilities in
tfx-bsl
(tfx-bsl>=0.15.2
) - Added documentation for Fairness Indicators.
- Deprecated component_type in favor of type.
- Deprecated component_id in favor of id.
- Move beam_pipeline_args out of additional_pipeline_args as top level pipeline param
- Deprecated chicago_taxi folder, beam setup scripts and serving examples are moved to chicago_taxi_pipeline folder.
- Moved beam setup scripts from examples/chicago_taxi/ to examples/chicago_taxi_pipeline/
- Moved interactive notebook classes into
tfx.orchestration.experimental
namespace. - Starting from 1.15, package
tensorflow
comes with GPU support. Users won't need to choose betweentensorflow
andtensorflow-gpu
. If any GPU devices are available, processes spawned by all TFX components will try to utilize them; note that in rare cases, this may exhaust the memory of the device(s). - Caveat:
tensorflow
2.0.0 is an exception and does not have GPU support. Iftensorflow-gpu
2.0.0 is installed before installingtfx
, it will be replaced withtensorflow
2.0.0. Re-installtensorflow-gpu
2.0.0 if needed. - Caveat: MLMD schema auto-upgrade is now disabled by default. For users who upgrades from 0.13 and do not want to lose the data in MLMD, please refer to MLMD documentation for guide to upgrade or downgrade MLMD database. Users who upgraded from TFX 0.14 should not be affected since there is not schema change between these two versions.
- Deprecated the usage of
tf.contrib.training.HParams
in Trainer as it is deprecated in TF 2.0. User module relying on member method of that class will not be supported. Dot style property access will be the only supported style from now on. - Any SavedModel produced by tf.Transform <=0.14 using any tf.contrib ops (or tf.Transform ops that used tf.contrib ops such as tft.quantiles, tft.bucketize, etc.) cannot be loaded with TF 2.0 since the contrib library has been removed in 2.0. Please refer to this issue.
- N/A
- Added conceptual info on Artifacts to guide/index.md
- Added support for Google Cloud ML Engine Training and Serving as extension.
- Supported pre-split input for ExampleGen components
- Added ImportExampleGen component for importing tfrecord files with TF Example data format
- Added a generic ExampleGen component to reduce the work of custom ExampleGen
- Released Python 3 type hints and added support for Python 3.6 and 3.7.
- Added an Airflow integration test for chicago_taxi_simple example.
- Updated tfx docker image to use Python 3.6 on Ubuntu 16.04.
- Added example for how to define and add a custom component.
- Added PrestoExampleGen component.
- Added Parquet executor for ExampleGen component.
- Added Avro executor for ExampleGen component.
- Enables Kubeflow Pipelines users to specify arbitrary ContainerOp decorators that can be applied to each pipeline step.
- Added scripts and instructions for running the TFX Chicago Taxi example on Spark (via Apache Beam).
- Introduced a new mechanism of artifact info passing between components that relies solely on ML Metadata.
- Unified driver and execution logging to go through tf.logging.
- Added support for Beam as an orchestrator.
- Introduced the experimental InteractiveContext environment for iterative notebook development, as well as an example Chicago Taxi notebook in this environment with TFDV / TFMA examples.
- Enabled Transform and Trainer components to specify user defined function (UDF) module by Python module path in addition to path to a module file.
- Enable ImportExampleGen component for Kubeflow.
- Enabled SchemaGen to infer feature shape.
- Enabled metadata logging and pipeline caching capability for KubeflowRunner.
- Used custom container for AI Platform Trainer extension.
- Introduced ExecutorSpec, which generalizes the representation of executors to include both Python classes and containers.
- Supported run context for metadata tracking of tfx pipeline.
- Deprecated 'metadata_db_root' in favor of passing in metadata_connection_config directly.
- airflow_runner.AirflowDAGRunner is renamed to airflow_dag_runner.AirflowDagRunner.
- runner.KubeflowRunner is renamed to kubeflow_dag_runner.KubeflowDagRunner.
- The "input" and "output" exec_properties fields for ExampleGen executors have been renamed to "input_config" and "output_config", respectively.
- Declared 'cmle_training_args' on trainer and 'cmle_serving_args' on pusher
deprecated. User should use the
trainer/pusher
executors in tfx.extensions.google_cloud_ai_platform module instead. - Moved tfx.orchestration.gcp.cmle_runner to tfx.extensions.google_cloud_ai_platform.runner.
- Deprecated csv_input and tfrecord_input, use external_input instead.
- Updated components and code samples to use
tft.TFTransformOutput
( introduced in tensorflow_transform 0.8). This avoids directly accessing the DatasetSchema object which may be removed in tensorflow_transform 0.14 or 0.15. - Fixed issue #113 to have consistent type of train_files and eval_files passed to trainer user module.
- Fixed issue #185 preventing the Airflow UI from visualizing the component's subdag operators and logs.
- Fixed issue #201 to make GCP credentials optional.
- Bumped dependency to kfp (Kubeflow Pipelines SDK) to be at version at least 0.1.18.
- Updated code example to
- use 'tf.data.TFRecordDataset' instead of the deprecated function 'tf.TFRecordReader'
- add test to train, evaluate and export.
- Component definition streamlined with explicit ComponentSpec and new style for defining component classes.
- TFX now depends on
pyarrow>=0.14.0,<0.15.0
(through its dependency ontensorflow-data-validation
). - Introduced 'examples' to the Trainer component API. It's recommended to use this field instead of 'transformed_examples' going forward.
- Trainer can now run without the 'transform_output' input.
- Added check for duplicated component ids within a pipeline.
- String representations for Channel and Artifact (TfxType) classes were improved.
- Updated workshop/setup/setup_demo.sh to fix version incompatibilities
- Updated workshop by adding note and instructions to fix issue with GCC
version when starting
airflow webserver
. - Prepared support for analyzer cache optimization in transform executor.
- Fixed issue #463 correcting syntax in SCHEMA_EMPTY message.
- Added an explicit check that pipeline name cannot exceed 63 characters.
- SchemaGen takes a new argument, infer_feature_shape to indicate whether to infer shape of features in schema. Current default value is False, but we plan to remove default value for it in future.
- Depended on 'click>=7.0,<8'
- Depended on
apache-beam[gcp]>=2.14,<3
- Depended on
ml-metadata>=-1.14.0,<0.15
- Depended on
tensorflow-data-validation>=0.14.1,<0.15
- Depended on
tensorflow-model-analysis>=0.14.0,<0.15
- Depended on
tensorflow-transform>=0.14.0,<0.15
- The "outputs" argument, which is used to override the automatically- generated output Channels for each component class has been removed; the equivalent overriding functionality is now available by specifying optional keyword arguments (see each component class definition for details).
- The optional arguments "executor" and "unique_name" of component classes have been uniformly renamed to "executor_spec" and "instance_name", respectively.
- The "driver" optional argument of component classes is no longer available: users who need to override the driver for a component should subclass the component and override the DRIVER_CLASS field.
- The
example_gen.component.ExampleGen
class has been refactored into theexample_gen.component._QueryBasedExampleGen
andexample_gen.component.FileBasedExampleGen
classes. - pipeline_root passed to pipeline.Pipeline is now the root to the running pipeline instead of root of all pipelines.
- Component class definitions have been simplified; existing custom components
need to:
- specify a ComponentSpec contract and conform to new class definition
style (see
base_component.BaseComponent
) - specify
EXECUTOR_SPEC=ExecutorClassSpec(MyExecutor)
in the component definition to replaceexecutor=MyExecutor
from component constructor.
- specify a ComponentSpec contract and conform to new class definition
style (see
- Artifact definitions for standard TFX components have moved from using
string type names into being concrete Artifact classes (see each official
TFX component's ComponentSpec definition in
types.standard_component_specs
and the definition of built-in Artifact types intypes.standard_artifacts
). - The
base_component.ComponentOutputs
class has been renamed tobase_component._PropertyDictWrapper
. - The tfx.utils.types.TfxType class has been renamed to tfx.types.Artifact.
- The tfx.utils.channel.Channel class has been moved to tfx.types.Channel.
- The "static_artifact_collection" argument to types.Channel has been renamed to "artifacts".
- ArtifactType for artifacts will have two new properties: pipeline_name and producer_component.
- The ARTIFACT_STATE_* constants were consolidated into the types.artifacts.ArtifactState enum class.
- Adds support for Python 3.5
- Initial version of following orchestration platform supported:
- Kubeflow
- Added TensorFlow Model Analysis Colab example
- Supported split ratio for ExampleGen components
- Supported running a single executor independently
- Fixes issue #43 that prevent new execution in some scenarios
- Fixes issue #47 that causes ImportError on chicago_taxi execution on dataflow
- Depends on
apache-beam[gcp]>=2.12,<3
- Depends on
tensorflow-data-validation>=0.13.1,<0.14
- Depends on
tensorflow-model-analysis>=0.13.2,<0.14
- Depends on
tensorflow-transform>=0.13,<0.14
- Deprecations:
- PipelineDecorator is deprecated. Please construct a pipeline directly from a list of components instead.
- Increased verbosity of logging to container stdout when running under Kubeflow Pipelines.
- Updated developer tutorial to support Python 3.5+
- Examples code are moved from 'examples' to 'tfx/examples': this ensures that PyPi package contains only one top level python module 'tfx'.
- N/A
- N/A
- Multiprocessing on Mac OS >= 10.13 might crash for Airflow. See AIRFLOW-3326 for details and solution.
- Adding TFMA Architecture doc
- TFX User Guide
- Initial version of the following TFX components:
- CSVExampleGen - CSV data ingestion
- BigQueryExampleGen - BigQuery data ingestion
- StatisticsGen - calculates statistics for the dataset
- SchemaGen - examines the dataset and creates a data schema
- ExampleValidator - looks for anomalies and missing values in the dataset
- Transform - performs feature engineering on the dataset
- Trainer - trains the model
- Evaluator - performs analysis of the model performance
- ModelValidator - helps validate exported models ensuring that they are "good enough" to be pushed to production
- Pusher - deploys the model to a serving infrastructure, for example the TensorFlow Serving Model Server
- Initial version of following orchestration platform supported:
- Apache Airflow
- Polished examples based on the Chicago Taxi dataset.
- Cleanup Colabs to remove TF warnings
- Performance improvement during shuffling of post-transform data.
- Changing example to move everything to one file in plugins
- Adding instructions to refer to README when running Chicago Taxi notebooks
- N/A
- N/A