@@ -50,6 +50,7 @@ def id(self):
5050
5151
5252class _CamelCaseMixin (BaseModel ):
53+
5354 class Config :
5455 allow_population_by_field_name = True
5556 alias_generator = camel_case
@@ -157,7 +158,7 @@ def _make_name_index(fields: List[DataRowMetadataSchema]):
157158
158159 @staticmethod
159160 def _make_id_index (
160- fields : List [DataRowMetadataSchema ]
161+ fields : List [DataRowMetadataSchema ]
161162 ) -> Dict [SchemaId , DataRowMetadataSchema ]:
162163 index = {}
163164 for f in fields :
@@ -194,23 +195,21 @@ def _parse_ontology(self) -> List[DataRowMetadataSchema]:
194195 for option in schema ["options" ]:
195196 option ["uid" ] = option .pop ("id" )
196197 options .append (
197- DataRowMetadataSchema (
198+ DataRowMetadataSchema (** {
199+ ** option ,
198200 ** {
199- ** option ,
200- ** {
201- "parent" : schema ["id" ]
202- }
203- })
204- )
201+ "parent" : schema ["id" ]
202+ }
203+ }))
205204 schema ["options" ] = options
206205 fields .append (DataRowMetadataSchema (** schema ))
207206
208207 return fields
209208
210209 def parse_metadata (
211- self , unparsed : List [Dict [str ,
212- List [Union [str ,
213- Dict ]]]]) -> List [DataRowMetadata ]:
210+ self , unparsed : List [Dict [str ,
211+ List [Union [str ,
212+ Dict ]]]]) -> List [DataRowMetadata ]:
214213 """ Parse metadata responses
215214
216215 >>> mdo.parse_metadata([metdata])
@@ -271,7 +270,7 @@ def bulk_upsert(
271270 raise ValueError ("Empty list passed" )
272271
273272 def _batch_upsert (
274- upserts : List [_UpsertBatchDataRowMetadata ]
273+ upserts : List [_UpsertBatchDataRowMetadata ]
275274 ) -> List [DataRowMetadataBatchResponse ]:
276275 query = """mutation UpsertDataRowMetadataBetaPyApi($metadata: [DataRowCustomMetadataBatchUpsertInput!]!) {
277276 upsertDataRowCustomMetadata(data: $metadata){
@@ -295,20 +294,22 @@ def _batch_upsert(
295294 items = []
296295 for m in metadata :
297296 if len (m .fields ) > _MAX_METADATA_FIELDS :
298- raise ValueError (f"Cannot upload { len (m .fields )} , the max number is { _MAX_METADATA_FIELDS } " )
297+ raise ValueError (
298+ f"Cannot upload { len (m .fields )} , the max number is { _MAX_METADATA_FIELDS } "
299+ )
299300 items .append (
300301 _UpsertBatchDataRowMetadata (
301302 data_row_id = m .data_row_id ,
302303 fields = list (
303304 chain .from_iterable (
304305 self ._parse_upsert (m ) for m in m .fields ))).dict (
305- by_alias = True ))
306+ by_alias = True ))
306307
307308 res = _batch_operations (_batch_upsert , items , self ._batch_size )
308309 return res
309310
310311 def bulk_delete (
311- self , deletes : List [DeleteDataRowMetadata ]
312+ self , deletes : List [DeleteDataRowMetadata ]
312313 ) -> List [DataRowMetadataBatchResponse ]:
313314 """ Delete metadata from a datarow by specifiying the fields you want to remove
314315
@@ -335,7 +336,7 @@ def bulk_delete(
335336 raise ValueError ("Empty list passed" )
336337
337338 def _batch_delete (
338- deletes : List [_DeleteBatchDataRowMetadata ]
339+ deletes : List [_DeleteBatchDataRowMetadata ]
339340 ) -> List [DataRowMetadataBatchResponse ]:
340341 query = """mutation DeleteDataRowMetadataBetaPyApi($deletes: [DataRowCustomMetadataBatchDeleteInput!]!) {
341342 deleteDataRowCustomMetadata(data: $deletes) {
@@ -454,9 +455,9 @@ def _batch_items(iterable: List[Any], size: int) -> Generator[Any, None, None]:
454455
455456
456457def _batch_operations (
457- batch_function : _BatchFunction ,
458- items : List ,
459- batch_size : int = 100 ,
458+ batch_function : _BatchFunction ,
459+ items : List ,
460+ batch_size : int = 100 ,
460461):
461462 response = []
462463
0 commit comments