@@ -793,7 +793,7 @@ async def test_store_rows_as_csv_being_split_and_uploaded(
793793 assert len (results ) == 200
794794
795795 # AND The spy should have been called in multiple batches
796- assert spy_send_job .call_count == 4
796+ assert spy_send_job .call_count == 5
797797
798798 async def test_store_rows_as_df_being_split_and_uploaded (
799799 self , project_model : Project , mocker : MockerFixture
@@ -857,7 +857,7 @@ async def test_store_rows_as_df_being_split_and_uploaded(
857857
858858 # AND The spy should have been called in multiple batches
859859 # Note: DataFrames have a minimum of 100 rows per batch
860- assert spy_send_job .call_count == 2
860+ assert spy_send_job .call_count == 3
861861
862862 @skip ("Skip in normal testing because the large size makes it slow" )
863863 async def test_store_rows_as_large_df_being_split_and_uploaded (
@@ -1040,7 +1040,7 @@ async def test_upsert_operations_with_various_data_sources(
10401040 # We should have 9 total rows now (6 from before + 3 new)
10411041 assert len (results ) == 9
10421042 # The spy should have been called for update and insert operations
1043- assert spy_send_job .call_count == 2
1043+ assert spy_send_job .call_count == 4
10441044
10451045 # Test 4: Dry run operation
10461046 # WHEN I perform a dry run upsert
@@ -1238,7 +1238,7 @@ async def test_upsert_with_large_data_and_batching(
12381238 assert len (results ) == 6
12391239
12401240 # AND multiple batch jobs should have been created due to batching settings
1241- assert spy_send_job .call_count == 5 # More batches due to small size settings
1241+ assert spy_send_job .call_count == 7 # More batches due to small size settings
12421242
12431243 async def test_upsert_all_data_types (
12441244 self , mocker : MockerFixture , project_model : Project
0 commit comments