Release Notes¶
v0.13.0 Feb 16, 2022¶
Warning
Woodwork may not support Python 3.7 in next non-bugfix release.
- Testing Changes
 
Replace mock with unittest.mock (#1304)
Thanks to the following people for contributing to this release: @dvreed77, @gsheni, @jeff-hernandez, @rwedge, @tamargrey, @thehomebrewnerd
v0.12.0 Jan 27, 2022¶
- Enhancements
 
Add Slack link to GitHub issue creation templates (#1242)
Thanks to the following people for contributing to this release: @dvreed77, @gsheni, @jeff-hernandez, @rwedge, @tamargrey, @thehomebrewnerd
v0.11.2 Jan 28, 2022¶
- Fixes
 
Set high and low bounds to the max and min values if no outliers are present in
box_plot_dict(backport of #1269)Thanks to the following people for contributing to this release: @tamargrey
Note¶
The pandas version for Koalas has been restricted, and a change was made to a pandas
replacecall to account for the recent pandas 1.4.0 release.
v0.11.1 Jan 4, 2022¶
- Changes
 
Update inference process to only check for NaturalLanguage if no other type matches are found first (#1234)
- Documentation Changes
 
Updating contributing doc with Spark installation instructions (#1232)
Thanks to the following people for contributing to this release: @gsheni, @thehomebrewnerd, @willsmithorg
v0.11.0 Dec 22, 2021¶
Thanks to the following people for contributing to this release: @jeff-hernandez, @thehomebrewnerd, @tuethan1999
v0.10.0 Nov 30, 2021¶
- Changes
 
Return histogram bins as a list of floats instead of a
pandas.Intervalobject (#1207)Thanks to the following people for contributing to this release: @tamargrey, @thehomebrewnerd
Breaking Changes¶
:pr:
1207: The behavior ofdescribe_dicthas changed when usingextra_stats=True. Previously, the histogram bins were returned aspandas.Intervalobjects. This has been updated so that the histogram bins are now represented as a two-element list of floats with the first element being the left edge of the bin and the second element being the right edge.
v0.9.1 Nov 19, 2021¶
- Fixes
 
Fix bug that causes
mutual_informationto fail with certain index types (#1199)
- Changes
 
Update pip to 21.3.1 for test requirements (#1196)
- Documentation Changes
 
Update install page with updated minimum optional dependencies (#1193)
Thanks to the following people for contributing to this release: @gsheni, @thehomebrewnerd
v0.9.0 Nov 11, 2021¶
- Enhancements
 
Added
read_fileparameter for replacing empty string values with NaN values (#1161)
- Changes
 
Make box plot low/high indices/values optional to return in
box_plot_dict(#1184)
- Documentation Changes
 
Update docs dependencies (#1176)
Thanks to the following people for contributing to this release: @bchen1116, @davesque, @gsheni, @rwedge, @tamargrey, @thehomebrewnerd
v0.8.2 Oct 12, 2021¶
Thanks to the following people for contributing to this release: @bchen1116, @davesque, @jeff-hernandez, @thehomebrewnerd
v0.8.1 Sep 16, 2021¶
- Changes
 
Update
Datetime.transformto use default nrows value when calling_infer_datetime_format(#1137)
- Documentation Changes
 
Hide spark config in Using Dask and Koalas Guide (#1139)
Thanks to the following people for contributing to this release: @jeff-hernandez, @simha104, @thehomebrewnerd
v0.8.0 Sep 9, 2021¶
- Changes
 
Update
get_invalid_schema_messageto improve performance (#1132)Thanks to the following people for contributing to this release: @ajaypallekonda, @davesque, @jeff-hernandez, @thehomebrewnerd
v0.7.1 Aug 25, 2021¶
Thanks to the following people for contributing to this release: @tamargrey
v0.7.0 Aug 25, 2021¶
Thanks to the following people for contributing to this release: @davesque, @jeff-hernandez, @tamargrey, @tuethan1999
Breaking Changes¶
:pr:
1100: The behavior forinithas changed. A full schema is a schema that contains all of the columns of the dataframe it describes whereas a partial schema only contains a subset. A full schema will also require that the schema is valid without having to make any changes to the DataFrame. Before, only a full schema was permitted by theinitmethod so passing a partial schema would error. Additionally, any parameters likelogical_typeswould be ignored if passing in a schema. Now, passing a partial schema to theinitmethod calls theinit_with_partial_schemamethod instead of throwing an error. Information from keyword arguments will override information from the partial schema. For example, if columnahas the Integer Logical Type in the partial schema, it’s possible to use thelogical_typeargument to reinfer it’s logical type by passing{'a': None}or force a type by passing in{'a': Double}. These changes mean that Woodwork init is less restrictive. If no type inference takes place and no changes are required of the DataFrame at initialization,init_with_full_schemashould be used instead ofinit.init_with_full_schemamaintains the same functionality as when a schema was passed to the oldinit.
v0.6.0 Aug 4, 2021¶
- Fixes
 
Fix bug in
_infer_datetime_formatwith allnp.naninput (#1089)
- Changes
 
The criteria for categorical type inference have changed (#1065)
The meaning of both the
categorical_thresholdandnumeric_categorical_thresholdsettings have changed (#1065)Make sampling for type inference more consistent (#1083)
Accessor logic checking if Woodwork has been initialized moved to decorator (#1093)
Thanks to the following people for contributing to this release: @davesque, @gsheni, @jeff-hernandez, @rwedge, @tamargrey, @thehomebrewnerd, @tuethan1999
Breaking Changes¶
#1065: The criteria for categorical type inference have changed. Relatedly, the meaning of both the
categorical_thresholdandnumeric_categorical_thresholdsettings have changed. Now, a categorical match is signaled when a series either has the “categorical” pandas dtype or if the ratio of unique value count (nan excluded) and total value count (nan also excluded) is below or equal to some fraction. The value used for this fraction is set by thecategorical_thresholdsetting which now has a default value of0.2. If a fraction is set for thenumeric_categorical_thresholdsetting, then series with either a float or integer dtype may be inferred as categorical by applying the same logic described above with thenumeric_categorical_thresholdfraction. Otherwise, thenumeric_categorical_thresholdsetting defaults toNonewhich indicates that series with a numerical type should not be inferred as categorical. Users who have overridden either thecategorical_thresholdornumeric_categorical_thresholdsettings will need to adjust their settings accordingly.#1083: The process of sampling series for logical type inference was updated to be more consistent. Before, initial sampling for inference differed depending on collection type (pandas, dask, or koalas). Also, further randomized subsampling was performed in some cases during categorical inference and in every case during email inference regardless of collection type. Overall, the way sampling was done was inconsistent and unpredictable. Now, the first 100,000 records of a column are sampled for logical type inference regardless of collection type although only records from the first partition of a dask dataset will be used. Subsampling performed by the inference functions of individual types has been removed. The effect of these changes is that inferred types may now be different although in many cases they will be more correct.
v0.5.1 Jul 22, 2021¶
- Enhancements
 
Store inferred datetime format on Datetime logical type instance (#1025)
Add support for automatically inferring the
EmailAddresslogical type (#1047)Add feature origin attribute to schema (#1056)
Add ability to calculate outliers and the statistical info required for box and whisker plots to
WoodworkColumnAccessor(#1048)Add ability to change config settings in a with block with
ww.config.with_options(#1062)
- Fixes
 
Raises warning and removes tags when user adds a column with index tags to DataFrame (#1035)
- Documentation Changes
 
Add supported types to metadata docstring (#1049)
Thanks to the following people for contributing to this release: @davesque, @frances-h, @jeff-hernandez, @simha104, @tamargrey, @thehomebrewnerd
v0.5.0 Jul 7, 2021¶
- Fixes
 
Add input validation to ww.init_series (#1015)
Thanks to the following people for contributing to this release: @jeff-hernandez, @simha104, @tamargrey, @thehomebrewnerd, @tuethan1999
Breaking Changes¶
The default logical type is now the
Unknowntype instead of theNaturalLanguagetype. The global confignatural_language_thresholdhas been renamed tocategorical_threshold.
v0.4.2 Jun 23, 2021¶
Thanks to the following people for contributing to this release: @frances-h, @gsheni, @jeff-hernandez, @tamargrey, @thehomebrewnerd, @tuethan1999
Breaking Changes¶
Progress callback functions parameters have changed and progress is now being reported in the units specified by the unit of measurement parameter instead of percentage of total. Progress callback functions now are expected to accept the following five parameters:
progress increment since last call
progress units complete so far
total units to complete
the progress unit of measurement
time elapsed since start of calculation
DataFrame.ww.initno longer accepts the make_index parameter
v0.4.1 Jun 9, 2021¶
- Changes
 
Change write_dataframe to only copy dataframe if it contains LatLong (#955)
Thanks to the following people for contributing to this release: @frances-h, @gsheni, @tamargrey, @thehomebrewnerd
v0.4.0 May 26, 2021¶
- Changes
 
Add logical type method to transform data (#915)
Thanks to the following people for contributing to this release: @gsheni, @jeff-hernandez, @thehomebrewnerd, @tuethan1999
v0.3.1 May 12, 2021¶
Warning
This Woodwork release uses a weak reference for maintaining a reference from the accessor to the DataFrame. Because of this, chaining a Woodwork call onto another call that creates a new DataFrame or Series object can be problematic.
Instead of calling
pd.DataFrame({'id':[1, 2, 3]}).ww.init(), first store the DataFrame in a new variable and then initialize Woodwork:df = pd.DataFrame({'id':[1, 2, 3]}) df.ww.init()
Thanks to the following people for contributing to this release: @gsheni, @jeff-hernandez, @tamargrey, @thehomebrewnerd
Breaking Changes¶
The command line functions
python -m woodwork list-logical-typesandpython -m woodwork list-semantic-tagsno longer exist. Please call the underlying Python functionsww.list_logical_types()andww.list_semantic_tags().
v0.3.0 May 3, 2021¶
- Enhancements
 
Add
is_schema_validandget_invalid_schema_messagefunctions for checking schema validity (#834)Add logical type for
AgeandAgeNullable(#849)Add logical type for
Address(#858)Add generic
to_diskfunction to save Woodwork schema and data (#872)Add generic
read_filefunction to read file as Woodwork DataFrame (#878)
- Changes
 
Consistently use
ColumnNotPresentErrorfor mismatches between user input and dataframe/schema columns (#837)Raise custom
WoodworkNotInitErrorwhen accessing Woodwork attributes before initialization (#838)Remove check requiring
Ordinalinstance for initializing aColumnSchemaobject (#870)Increase koalas min version to 1.8.0 (#885)
- Documentation Changes
 
Improve formatting of release notes (#874)
Thanks to the following people for contributing to this release: @frances-h, @gsheni, @jeff-hernandez, @rwedge, @tamargrey, @thehomebrewnerd
Breaking Changes¶
Woodwork tables can no longer be saved using to disk
df.ww.to_csv,df.ww.to_pickle, ordf.ww.to_parquet. Usedf.ww.to_diskinstead.The
read_csvfunction has been replaced byread_file.
v0.2.0 Apr 20, 2021¶
Warning
This Woodwork release does not support Python 3.6
- Enhancements
 
Add validation control to WoodworkTableAccessor (#736)
Store
make_indexvalue on WoodworkTableAccessor (#780)Add optional
excludeparameter to WoodworkTableAccessorselectmethod (#783)Add validation control to
deserialize.read_woodwork_tableandww.read_csv(#788)Add
WoodworkColumnAccessor.schemaand handle copying column schema (#799)Allow initializing a
WoodworkColumnAccessorwith aColumnSchema(#814)Add
__repr__toColumnSchema(#817)Add
BooleanNullableandIntegerNullablelogical types (#830)Add validation control to
WoodworkColumnAccessor(#833)
- Changes
 
Rename
FullNamelogical type toPersonFullName(#740)Rename
ZIPCodelogical type toPostalCode(#741)Update minimum scikit-learn version to 0.22 (#763)
Drop support for Python version 3.6 (#768)
Remove
ColumnNameMismatchWarning(#777)
get_column_dictdoes not use standard tags by default (#782)Make
logical_typeandnameparams to_get_column_dictoptional (#786)Rename Schema object and files to match new table-column schema structure (#789)
Store column typing information in a
ColumnSchemaobject instead of a dictionary (#791)
TableSchemadoes not use standard tags by default (#806)Store
use_standard_tagson theColumnSchemainstead of theTableSchema(#809)Move functions in
column_schema.pyto be methods onColumnSchema(#829)
- Testing Changes
 
Add unit tests against minimum dependencies for python 3.6 on PRs and main (#743, #753, #763)
Update spark config for test fixtures (#787)
Separate latest unit tests into pandas, dask, koalas (#813)
Update latest dependency checker to generate separate core, koalas, and dask dependencies (#815, #825)
Ignore latest dependency branch when checking for updates to the release notes (#827)
Change from GitHub PAT to auto generated GitHub Token for dependency checker (#831)
Expand
ColumnSchemasemantic tag testing coverage and nulllogical_typetesting coverage (#832)Thanks to the following people for contributing to this release: @gsheni, @jeff-hernandez, @rwedge, @tamargrey, @thehomebrewnerd
Breaking Changes¶
The
ZIPCodelogical type has been renamed toPostalCodeThe
FullNamelogical type has been renamed toPersonFullNameThe
Schemaobject has been renamed toTableSchemaWith the
ColumnSchemaobject, typing information for a column can no longer be accessed withdf.ww.columns[col_name]['logical_type']. Instead usedf.ww.columns[col_name].logical_type.The
BooleanandIntegerlogical types will no longer work with data that contains null values. The newBooleanNullableandIntegerNullablelogical types should be used if null values are present.
v0.1.0 Mar 22, 2021¶
- Enhancements
 
Implement Schema and Accessor API (#497)
Add Schema class that holds typing info (#499)
Add WoodworkTableAccessor class that performs type inference and stores Schema (#514)
Allow initializing Accessor schema with a valid Schema object (#522)
Add ability to read in a csv and create a DataFrame with an initialized Woodwork Schema (#534)
Add ability to call pandas methods from Accessor (#538, #589)
Add helpers for checking if a column is one of Boolean, Datetime, numeric, or categorical (#553)
Add ability to load demo retail dataset with a Woodwork Accessor (#556)
Add
selectto WoodworkTableAccessor (#548)Add
mutual_informationto WoodworkTableAccessor (#571)Add WoodworkColumnAccessor class (#562)
Add semantic tag update methods to column accessor (#573)
Add
describeanddescribe_dictto WoodworkTableAccessor (#579)Add
init_seriesutil function for initializing a series with dtype change (#581)Add
set_logical_typemethod to WoodworkColumnAccessor (#590)Add semantic tag update methods to table schema (#591)
Add warning if additional parameters are passed along with schema (#593)
Better warning when accessing column properties before init (#596)
Update column accessor to work with LatLong columns (#598)
Add
set_indexto WoodworkTableAccessor (#603)Implement
locandilocfor WoodworkColumnAccessor (#613)Add
set_time_indexto WoodworkTableAccessor (#612)Implement
locandilocfor WoodworkTableAccessor (#618)Allow updating logical types with
set_typesand make relevant DataFrame changes (#619)Allow serialization of WoodworkColumnAccessor to csv, pickle, and parquet (#624)
Add DaskColumnAccessor (#625)
Allow deserialization from csv, pickle, and parquet to Woodwork table (#626)
Add
value_countsto WoodworkTableAccessor (#632)Add KoalasColumnAccessor (#634)
Add
popto WoodworkTableAccessor (#636)Add
dropto WoodworkTableAccessor (#640)Add
renameto WoodworkTableAccessor (#646)Add DaskTableAccessor (#648)
Add Schema properties to WoodworkTableAccessor (#651)
Add KoalasTableAccessor (#652)
Adds
__getitem__to WoodworkTableAccessor (#633)Update Koalas min version and add support for more new pandas dtypes with Koalas (#678)
Adds
__setitem__to WoodworkTableAccessor (#669)
- Changes
 
Move mutual information logic to statistics utils file (#584)
Bump min Koalas version to 1.4.0 (#638)
Preserve pandas underlying index when not creating a Woodwork index (#664)
Restrict Koalas version to
<1.7.0due to breaking changes (#674)Clean up dtype usage across Woodwork (#682)
Improve error when calling accessor properties or methods before init (#683)
Remove dtype from Schema dictionary (#685)
Add
include_indexparam and allow unique columns in Accessor mutual information (#699)Include DataFrame equality and
use_standard_tagsin WoodworkTableAccessor equality check (#700)Remove
DataTableandDataColumnclasses to migrate towards the accessor approach (#713)Change
sample_seriesdtype to not need conversion and removeconvert_seriesutil (#720)Rename Accessor methods since
DataTablehas been removed (#723)
- Documentation Changes
 
Update README.md and Get Started guide to use accessor (#655, #717)
Update Understanding Types and Tags guide to use accessor (#657)
Update docstrings and API Reference page (#660)
Update statistical insights guide to use accessor (#693)
Update Customizing Type Inference guide to use accessor (#696)
Update Dask and Koalas guide to use accessor (#701)
Update index notebook and install guide to use accessor (#715)
Add section to documentation about schema validity (#729)
Update README.md and Get Started guide to use
pd.read_csv(#730)Make small fixes to documentation formatting (#731)
Thanks to the following people for contributing to this release: @gsheni, @jeff-hernandez, @johnbridstrup, @tamargrey, @thehomebrewnerd
Breaking Changes¶
The
DataTableandDataColumnclasses have been removed and replaced by newWoodworkTableAccessorandWoodworkColumnAccessorclasses which are used through thewwnamespace available on DataFrames after importing Woodwork.
v0.0.11 Mar 15, 2021¶
- Documentation Changes
 
Update to remove warning message from statistical insights guide (#690)
Thanks to the following people for contributing to this release: @gsheni, @tamargrey, @thehomebrewnerd
v0.0.10 Feb 25, 2021¶
Thanks to the following people for contributing to this release: @gsheni, @jeff-hernandez, @johnbridstrup, @tamargrey
v0.0.9 Feb 5, 2021¶
Thanks to the following people for contributing to this release: @gsheni, @tamargrey, @thehomebrewnerd
v0.0.8 Jan 25, 2021¶
Thanks to the following people for contributing to this release: @gsheni, @tamargrey, @thehomebrewnerd
v0.0.7 Dec 14, 2020¶
- Changes
 
Update links to use alteryx org Github URL (#423)
Support column names of any type allowed by the underlying DataFrame (#442)
Use
objectdtype for LatLong columns for easy access to latitude and longitude values (#414)Restrict dask version to prevent 2020.12.0 release from being installed (#453)
Lower minimum requirement for numpy to 1.15.4, and set pandas minimum requirement 1.1.1 (#459)
- Testing Changes
 
Fix missing test coverage (#436)
Thanks to the following people for contributing to this release: @gsheni, @jeff-hernandez, @tamargrey, @thehomebrewnerd
v0.0.6 Nov 30, 2020¶
- Enhancements
 
Add support for creating DataTable from Koalas DataFrame (#327)
Add ability to initialize DataTable with numpy array (#367)
Add
describe_dictmethod to DataTable (#405)Add
mutual_information_dictmethod to DataTable (#404)Add
metadatato DataTable for user-defined metadata (#392)Add
update_dataframemethod to DataTable to update underlying DataFrame (#407)Sort dataframe if
time_indexis specified, bypass sorting withalready_sortedparameter. (#410)Add
descriptionattribute to DataColumn (#416)Implement
DataColumn.__len__andDataTable.__len__(#415)
- Changes
 
Lower moto test requirement for serialization/deserialization (#376)
Make Koalas an optional dependency installable with woodwork[koalas] (#378)
Remove WholeNumber LogicalType from Woodwork (#380)
Updates to LogicalTypes to support Koalas 1.4.0 (#393)
Replace
set_logical_typesandset_semantic_tagswith justset_types(#379)Remove
copy_dataframeparameter from DataTable initialization (#398)Implement
DataTable.__sizeof__to return size of the underlying dataframe (#401)Include Datetime columns in mutual info calculation (#399)
Maintain column order on DataTable operations (#406)
Thanks to the following people for contributing to this release: @ctduffy, @gsheni, @tamargrey, @thehomebrewnerd
Breaking Changes¶
The
DataTable.set_semantic_tagsmethod was removed.DataTable.set_typescan be used instead.The
DataTable.set_logical_typesmethod was removed.DataTable.set_typescan be used instead.
WholeNumberwas removed from LogicalTypes. Columns that were previously inferred as WholeNumber will now be inferred as Integer.The
DataTable.get_mutual_informationwas renamed toDataTable.mutual_information.The
copy_dataframeparameter was removed from DataTable initialization.
v0.0.5 Nov 11, 2020¶
- Enhancements
 
Add
__eq__to DataTable and DataColumn and update LogicalType equality (#318)Add
value_counts()method to DataTable (#342)Support serialization and deserialization of DataTables via csv, pickle, or parquet (#293)
Add
shapeproperty to DataTable and DataColumn (#358)Add
ilocmethod to DataTable and DataColumn (#365)Add
numeric_categorical_thresholdconfig value to allow inferring numeric columns as Categorical (#363)Add
renamemethod to DataTable (#367)
- Fixes
 
Catch non numeric time index at validation (#332)
- Changes
 
Support logical type inference from a Dask DataFrame (#248)
Fix validation checks and
make_indexto work with Dask DataFrames (#260)Skip validation of Ordinal order values for Dask DataFrames (#270)
Improve support for datetimes with Dask input (#286)
Update
DataTable.describeto work with Dask input (#296)Update
DataTable.get_mutual_informationto work with Dask input (#300)Modify
to_pandasfunction to return DataFrame with correct index (#281)Rename
DataColumn.to_pandasmethod toDataColumn.to_series(#311)Rename
DataTable.to_pandasmethod toDataTable.to_dataframe(#319)Remove UserWarning when no matching columns found (#325)
Remove
copyparameter fromDataTable.to_dataframeandDataColumn.to_series(#338)Allow pandas ExtensionArrays as inputs to DataColumn (#343)
Move warnings to a separate exceptions file and call via UserWarning subclasses (#348)
Make Dask an optional dependency installable with woodwork[dask] (#357)
Thanks to the following people for contributing to this release: @ctduffy, @gsheni, @tamargrey, @thehomebrewnerd
Breaking Changes¶
The
DataColumn.to_pandasmethod was renamed toDataColumn.to_series.The
DataTable.to_pandasmethod was renamed toDataTable.to_dataframe.
copyis no longer a parameter ofDataTable.to_dataframeorDataColumn.to_series.
v0.0.4 Oct 21, 2020¶
- Enhancements
 
Add optional
includeparameter forDataTable.describe()to filter results (#228)Add
make_indexparameter toDataTable.__init__to enable optional creation of a new index column (#238)Add support for setting ranking order on columns with Ordinal logical type (#240)
Add
list_semantic_tagsfunction and CLI to get dataframe of woodwork semantic_tags (#244)Add support for numeric time index on DataTable (#267)
Add pop method to DataTable (#289)
Add entry point to setup.py to run CLI commands (#285)
- Fixes
 
Allow numeric datetime time indices (#282)
Thanks to the following people for contributing to this release: @ctduffy, @gsheni, @tamargrey, @thehomebrewnerd
v0.0.3 Oct 9, 2020¶
- Enhancements
 
Implement setitem on DataTable to create/overwrite an existing DataColumn (#165)
Add
to_pandasmethod to DataColumn to access the underlying series (#169)Add list_logical_types function and CLI to get dataframe of woodwork LogicalTypes (#172)
Add
describemethod to DataTable to generate statistics for the underlying data (#181)Add optional
return_dataframeparameter toload_retailto return either DataFrame or DataTable (#189)Add
get_mutual_informationmethod to DataTable to generate mutual information between columns (#203)Add
read_csvfunction to create DataTable directly from CSV file (#222)
- Changes
 
Remove unnecessary
add_standard_tagsattribute from DataTable (#171)Remove standard tags from index column and do not return stats for index column from
DataTable.describe(#196)Update
DataColumn.set_semantic_tagsandDataColumn.add_semantic_tagsto return new objects (#205)Update various DataTable methods to return new objects rather than modifying in place (#210)
Move datetime_format to Datetime LogicalType (#216)
Do not calculate mutual info with index column in
DataTable.get_mutual_information(#221)Move setting of underlying physical types from DataTable to DataColumn (#233)
- Documentation Changes
 
Remove unused code from sphinx conf.py, update with Github URL(#160, #163)
Update README and docs with new Woodwork logo, with better code snippets (#161, #159)
Add DataTable and DataColumn to API Reference (#162)
Add docstrings to LogicalType classes (#168)
Add Woodwork image to index, clear outputs of Jupyter notebook in docs (#173)
Update contributing.md, release.md with all instructions (#176)
Add section for setting index and time index to start notebook (#179)
Rename changelog to Release Notes (#193)
Add section for standard tags to start notebook (#188)
Add Understanding Types and Tags user guide (#201)
Add missing docstring to
list_logical_types(#202)Add Woodwork Global Configuration Options guide (#215)
Thanks to the following people for contributing to this release: @gsheni, @tamargrey, @thehomebrewnerd
v0.0.2 Sep 28, 2020¶
- Fixes
 
Fix formatting issue when printing global config variables (#138)
- Documentation Changes
 
Add working code example to README and create Using Woodwork page (#103)
Thanks to the following people for contributing to this release: @gsheni, @tamargrey, @thehomebrewnerd
v0.1.0 Sep 24, 2020¶
Add
natural_language_thresholdglobal config option used for Categorical/NaturalLanguage type inference (#135)Add global config options and add
datetime_formatoption for type inference (#134)Fix bug with Integer and WholeNumber inference in column with
pd.NAvalues (#133)Add
DataTable.ltypesproperty to return series of logical types (#131)Add ability to create new datatable from specified columns with
dt[[columns]](#127)Handle setting and tagging of index and time index columns (#125)
Add combined tag and ltype selection (#124)
Add changelog, and update changelog check to CI (#123)
Implement
reset_semantic_tags(#118)Implement DataTable getitem (#119)
Add
remove_semantic_tagsmethod (#117)Add semantic tag selection (#106)
Add github action, rename to woodwork (#113)
Add license to setup.py (#112)
Reset semantic tags on logical type change (#107)
Add standard numeric and category tags (#100)
Change
semantic_typestosemantic_tags, a set of strings (#100)Update dataframe dtypes based on logical types (#94)
Add
select_logical_typesto DataTable (#96)Add pygments to dev-requirements.txt (#97)
Add replacing None with np.nan in DataTable init (#87)
Refactor DataColumn to make
semantic_typesandlogical_typeprivate (#86)Add pandas_dtype to each Logical Type, and remove dtype attribute on DataColumn (#85)
Add set_semantic_types methods on both DataTable and DataColumn (#75)
Support passing camel case or snake case strings for setting logical types (#74)
Improve flexibility when setting semantic types (#72)
Add Whole Number Inference of Logical Types (#66)
Add
dtypesproperty to DataTables andreprfor DataColumn (#61)Allow specification of semantic types during DataTable creation (#69)
Implements
set_logical_typeson DataTable (#65)Add init files to tests to fix code coverage (#60)
Add AutoAssign bot (#59)
Add logical types validation in DataTables (#49)
Fix working_directory in CI (#57)
Add
infer_logical_typesfor DataColumn (#45)Add code coverage (#51)
Improve and refactor the validation checks during initialization of a DataTable (#40)
Add dataframe attribute to DataTable (#39)
Update ReadME with minor usage details (#37)
Add License (#34)
Rename from datatables to datatables (#4)
Add Logical Types, DataTable, DataColumn (#3)
Add Makefile, setup.py, requirements.txt (#2)
Initial Release (#1)
Thanks to the following people for contributing to this release: @gsheni, @tamargrey, @thehomebrewnerd