test_data_node.py 36 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503504505506507508509510511512513514515516517518519520521522523524525526527528529530531532533534535536537538539540541542543544545546547548549550551552553554555556557558559560561562563564565566567568569570571572573574575576577578579580581582583584585586587588589590591592593594595596597598599600601602603604605606607608609610611612613614615616617618619620621622623624625626627628629630631632633634635636637638639640641642643644645646647648649650651652653654655656657658659660661662663664665666667668669670671672673674675676677678679680681682683684685686687688689690691692693694695696697698699700701702703704705706707708709710711712713714715716717718719720721722723724725726727728729730731732733734735736737738739740741742743744745746747748749750751752753754755756757758759760761762763764765766767768769770771772773774775776777778779780781782783784785786787788789790791792793794795796797798799800801802803804805806807808809810811812813814815816817818819820821822823824825826827828829830831832833834835836837838839840841842843844845846847848849850851852853854855856857858859860861862863864865866867868869870871872873874875876877878879880881882883884885886887888889890891892893894895896897898899900901902903904905906907908909910911912913914915916917918919920921922923924925926927928929930931932933934935936937938939940941942943944945946947948949950951952953
  1. # Copyright 2021-2025 Avaiga Private Limited
  2. #
  3. # Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
  4. # the License. You may obtain a copy of the License at
  5. #
  6. # http://www.apache.org/licenses/LICENSE-2.0
  7. #
  8. # Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on
  9. # an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the
  10. # specific language governing permissions and limitations under the License.
  11. import os
  12. from datetime import datetime, timedelta
  13. from time import sleep
  14. from unittest import mock
  15. import freezegun
  16. import pandas as pd
  17. import pytest
  18. import taipy.core as tp
  19. from taipy.common.config import Config
  20. from taipy.common.config.common.scope import Scope
  21. from taipy.common.config.exceptions.exceptions import InvalidConfigurationId
  22. from taipy.core.data._data_manager import _DataManager
  23. from taipy.core.data._data_manager_factory import _DataManagerFactory
  24. from taipy.core.data.data_node import DataNode
  25. from taipy.core.data.data_node_id import (
  26. EDIT_COMMENT_KEY,
  27. EDIT_EDITOR_ID_KEY,
  28. EDIT_JOB_ID_KEY,
  29. EDIT_TIMESTAMP_KEY,
  30. DataNodeId,
  31. )
  32. from taipy.core.data.in_memory import InMemoryDataNode
  33. from taipy.core.exceptions.exceptions import DataNodeIsBeingEdited, NoData
  34. from taipy.core.job.job_id import JobId
  35. from taipy.core.task.task import Task
  36. from .utils import FakeDataNode
  37. def funct_a_b(input: str):
  38. print("task_a_b") # noqa: T201
  39. return "B"
  40. def funct_b_c(input: str):
  41. print("task_b_c") # noqa: T201
  42. return "C"
  43. def funct_b_d(input: str):
  44. print("task_b_d") # noqa: T201
  45. return "D"
  46. class TestDataNode:
  47. def test_dn_equals(self, data_node):
  48. data_manager = _DataManagerFactory()._build_manager()
  49. dn_id = data_node.id
  50. data_manager._set(data_node)
  51. # # To test if instance is same type
  52. task = Task("task", {}, print, [], [], dn_id)
  53. dn_2 = data_manager._get(dn_id)
  54. assert data_node == dn_2
  55. assert data_node != dn_id
  56. assert data_node != task
  57. def test_create_with_default_values(self):
  58. dn = DataNode("foo_bar")
  59. assert dn.config_id == "foo_bar"
  60. assert dn.scope == Scope.SCENARIO
  61. assert dn.id is not None
  62. assert dn.name is None
  63. assert dn.owner_id is None
  64. assert dn.parent_ids == set()
  65. assert dn.last_edit_date is None
  66. assert dn.job_ids == []
  67. assert not dn.is_ready_for_reading
  68. assert len(dn.properties) == 0
  69. def test_create_with_ranks(self):
  70. # Test _rank is propagated from the config
  71. cfg = Config.configure_data_node("foo_bar")
  72. cfg._ranks = {"A": 1, "B": 2, "C": 0}
  73. dn = DataNode("foo_bar")
  74. assert dn.config_id == "foo_bar"
  75. assert dn.scope == Scope.SCENARIO
  76. assert dn.id is not None
  77. assert dn.name is None
  78. assert dn.owner_id is None
  79. assert dn.parent_ids == set()
  80. assert dn.last_edit_date is None
  81. assert dn.job_ids == []
  82. assert not dn.is_ready_for_reading
  83. assert len(dn.properties) == 0
  84. assert dn._get_rank("A") == 1
  85. assert dn._get_rank("B") == 2
  86. assert dn._get_rank("C") == 0
  87. def test_is_up_to_date_when_not_written(self):
  88. dn_confg_1 = Config.configure_in_memory_data_node("dn_1", default_data="a")
  89. dn_confg_2 = Config.configure_in_memory_data_node("dn_2")
  90. task_config_1 = Config.configure_task("t1", funct_a_b, [dn_confg_1], [dn_confg_2])
  91. scenario_config = Config.configure_scenario("sc", [task_config_1])
  92. scenario = tp.create_scenario(scenario_config)
  93. assert scenario.dn_1.is_up_to_date is True
  94. assert scenario.dn_2.is_up_to_date is False
  95. tp.submit(scenario)
  96. assert scenario.dn_1.is_up_to_date is True
  97. assert scenario.dn_2.is_up_to_date is True
  98. def test_create(self):
  99. a_date = datetime.now()
  100. dn = DataNode(
  101. "foo_bar",
  102. Scope.SCENARIO,
  103. DataNodeId("an_id"),
  104. "a_scenario_id",
  105. {"a_parent_id"},
  106. a_date,
  107. [{"job_id": "a_job_id"}],
  108. edit_in_progress=False,
  109. prop="erty",
  110. name="a name",
  111. )
  112. assert dn.config_id == "foo_bar"
  113. assert dn.scope == Scope.SCENARIO
  114. assert dn.id == "an_id"
  115. assert dn.name == "a name"
  116. assert dn.owner_id == "a_scenario_id"
  117. assert dn.parent_ids == {"a_parent_id"}
  118. assert dn.last_edit_date == a_date
  119. assert dn.job_ids == ["a_job_id"]
  120. assert dn.is_ready_for_reading
  121. assert len(dn.properties) == 2
  122. assert dn.properties == {"prop": "erty", "name": "a name"}
  123. with pytest.raises(InvalidConfigurationId):
  124. DataNode("foo bar")
  125. def test_read_write(self):
  126. dn = FakeDataNode("foo_bar")
  127. with pytest.raises(NoData):
  128. assert dn.read() is None
  129. dn.read_or_raise()
  130. assert dn.write_has_been_called == 0
  131. assert dn.read_has_been_called == 0
  132. assert not dn.is_ready_for_reading
  133. assert dn.last_edit_date is None
  134. assert dn.job_ids == []
  135. assert dn.edits == []
  136. dn.write("Any data")
  137. assert dn.write_has_been_called == 1
  138. assert dn.read_has_been_called == 0
  139. assert dn.last_edit_date is not None
  140. first_edition = dn.last_edit_date
  141. assert dn.is_ready_for_reading
  142. assert dn.job_ids == []
  143. assert len(dn.edits) == 1
  144. assert dn.get_last_edit()["timestamp"] == dn.last_edit_date
  145. sleep(0.1)
  146. dn.write("Any other data", job_id := JobId("a_job_id"))
  147. assert dn.write_has_been_called == 2
  148. assert dn.read_has_been_called == 0
  149. second_edition = dn.last_edit_date
  150. assert first_edition < second_edition
  151. assert dn.is_ready_for_reading
  152. assert dn.job_ids == [job_id]
  153. assert len(dn.edits) == 2
  154. assert dn.get_last_edit()["timestamp"] == dn.last_edit_date
  155. dn.read()
  156. assert dn.write_has_been_called == 2
  157. assert dn.read_has_been_called == 1
  158. second_edition = dn.last_edit_date
  159. assert first_edition < second_edition
  160. assert dn.is_ready_for_reading
  161. assert dn.job_ids == [job_id]
  162. def test_lock_initialization(self):
  163. dn = InMemoryDataNode("dn", Scope.SCENARIO)
  164. assert not dn.edit_in_progress
  165. assert dn._editor_id is None
  166. assert dn._editor_expiration_date is None
  167. def test_locked_dn_unlockable_only_by_same_editor(self):
  168. dn = InMemoryDataNode("dn", Scope.SCENARIO)
  169. dn.lock_edit("user_1")
  170. assert dn.edit_in_progress
  171. assert dn._editor_id == "user_1"
  172. assert dn._editor_expiration_date is not None
  173. with pytest.raises(DataNodeIsBeingEdited):
  174. dn.lock_edit("user_2")
  175. with pytest.raises(DataNodeIsBeingEdited):
  176. dn.unlock_edit("user_2")
  177. dn.unlock_edit("user_1")
  178. assert not dn.edit_in_progress
  179. assert dn._editor_id is None
  180. assert dn._editor_expiration_date is None
  181. def test_none_editor_can_lock_a_locked_dn(self):
  182. dn = InMemoryDataNode("dn", Scope.SCENARIO)
  183. dn.lock_edit("user")
  184. assert dn.edit_in_progress
  185. assert dn._editor_id == "user"
  186. assert dn._editor_expiration_date is not None
  187. dn.lock_edit()
  188. assert dn.edit_in_progress
  189. assert dn._editor_id is None
  190. assert dn._editor_expiration_date is None
  191. def test_none_editor_can_unlock_a_locked_dn(self):
  192. dn = InMemoryDataNode("dn", Scope.SCENARIO)
  193. dn.lock_edit("user")
  194. assert dn.edit_in_progress
  195. assert dn._editor_id == "user"
  196. assert dn._editor_expiration_date is not None
  197. dn.unlock_edit()
  198. assert not dn.edit_in_progress
  199. assert dn._editor_id is None
  200. assert dn._editor_expiration_date is None
  201. dn.lock_edit()
  202. assert dn.edit_in_progress
  203. assert dn._editor_id is None
  204. assert dn._editor_expiration_date is None
  205. dn.unlock_edit()
  206. assert not dn.edit_in_progress
  207. assert dn._editor_id is None
  208. assert dn._editor_expiration_date is None
  209. def test_ready_for_reading(self):
  210. dn = InMemoryDataNode("foo_bar", Scope.CYCLE)
  211. assert dn.last_edit_date is None
  212. assert not dn.is_ready_for_reading
  213. assert dn.job_ids == []
  214. dn.lock_edit()
  215. assert dn.last_edit_date is None
  216. assert not dn.is_ready_for_reading
  217. assert dn.job_ids == []
  218. dn.unlock_edit()
  219. assert dn.last_edit_date is None
  220. assert not dn.is_ready_for_reading
  221. assert dn.job_ids == []
  222. dn.lock_edit()
  223. assert dn.last_edit_date is None
  224. assert not dn.is_ready_for_reading
  225. assert dn.job_ids == []
  226. dn.write("toto", job_id := JobId("a_job_id"))
  227. assert dn.last_edit_date is not None
  228. assert dn.is_ready_for_reading
  229. assert dn.job_ids == [job_id]
  230. def test_is_valid_no_validity_period(self):
  231. # Test Never been written
  232. dn = InMemoryDataNode("foo", Scope.SCENARIO, DataNodeId("id"), "name", "owner_id")
  233. assert not dn.is_valid
  234. # test has been written
  235. dn.write("My data")
  236. assert dn.is_valid
  237. def test_is_valid_with_30_min_validity_period(self):
  238. # Test Never been written
  239. dn = InMemoryDataNode(
  240. "foo", Scope.SCENARIO, DataNodeId("id"), "name", "owner_id", validity_period=timedelta(minutes=30)
  241. )
  242. assert dn.is_valid is False
  243. # Has been written less than 30 minutes ago
  244. dn.write("My data")
  245. assert dn.is_valid is True
  246. # Has been written more than 30 minutes ago
  247. dn.last_edit_date = datetime.now() + timedelta(days=-1)
  248. assert dn.is_valid is False
  249. def test_is_valid_with_5_days_validity_period(self):
  250. # Test Never been written
  251. dn = InMemoryDataNode("foo", Scope.SCENARIO, validity_period=timedelta(days=5))
  252. assert dn.is_valid is False
  253. # Has been written less than 30 minutes ago
  254. dn.write("My data")
  255. assert dn.is_valid is True
  256. # Has been written more than 30 minutes ago
  257. dn._last_edit_date = datetime.now() - timedelta(days=6)
  258. _DataManager()._set(dn)
  259. assert dn.is_valid is False
  260. def test_is_up_to_date(self, current_datetime):
  261. dn_confg_1 = Config.configure_in_memory_data_node("dn_1")
  262. dn_confg_2 = Config.configure_in_memory_data_node("dn_2")
  263. dn_confg_3 = Config.configure_in_memory_data_node("dn_3", scope=Scope.GLOBAL)
  264. task_config_1 = Config.configure_task("t1", print, [dn_confg_1], [dn_confg_2])
  265. task_config_2 = Config.configure_task("t2", print, [dn_confg_2], [dn_confg_3])
  266. scenario_config = Config.configure_scenario("sc", [task_config_1, task_config_2])
  267. scenario_1 = tp.create_scenario(scenario_config)
  268. assert len(_DataManager._get_all()) == 3
  269. dn_1_1 = scenario_1.data_nodes["dn_1"]
  270. dn_2_1 = scenario_1.data_nodes["dn_2"]
  271. dn_3_1 = scenario_1.data_nodes["dn_3"]
  272. assert dn_1_1.last_edit_date is None
  273. assert dn_2_1.last_edit_date is None
  274. assert dn_3_1.last_edit_date is None
  275. dn_1_1.last_edit_date = current_datetime + timedelta(1)
  276. dn_2_1.last_edit_date = current_datetime + timedelta(2)
  277. dn_3_1.last_edit_date = current_datetime + timedelta(3)
  278. assert dn_1_1.is_up_to_date
  279. assert dn_2_1.is_up_to_date
  280. assert dn_3_1.is_up_to_date
  281. dn_2_1.last_edit_date = current_datetime + timedelta(4)
  282. assert dn_1_1.is_up_to_date
  283. assert dn_2_1.is_up_to_date
  284. assert not dn_3_1.is_up_to_date
  285. dn_1_1.last_edit_date = current_datetime + timedelta(5)
  286. assert dn_1_1.is_up_to_date
  287. assert not dn_2_1.is_up_to_date
  288. assert not dn_3_1.is_up_to_date
  289. dn_1_1.last_edit_date = current_datetime + timedelta(1)
  290. dn_2_1.last_edit_date = current_datetime + timedelta(2)
  291. dn_3_1.last_edit_date = current_datetime + timedelta(3)
  292. def test_is_up_to_date_across_scenarios(self, current_datetime):
  293. dn_confg_1 = Config.configure_in_memory_data_node("dn_1", scope=Scope.SCENARIO)
  294. dn_confg_2 = Config.configure_in_memory_data_node("dn_2", scope=Scope.SCENARIO)
  295. dn_confg_3 = Config.configure_in_memory_data_node("dn_3", scope=Scope.GLOBAL)
  296. task_config_1 = Config.configure_task("t1", print, [dn_confg_1], [dn_confg_2])
  297. task_config_2 = Config.configure_task("t2", print, [dn_confg_2], [dn_confg_3])
  298. scenario_config = Config.configure_scenario("sc", [task_config_1, task_config_2])
  299. scenario_1 = tp.create_scenario(scenario_config)
  300. scenario_2 = tp.create_scenario(scenario_config)
  301. assert len(_DataManager._get_all()) == 5
  302. dn_1_1 = scenario_1.data_nodes["dn_1"]
  303. dn_2_1 = scenario_1.data_nodes["dn_2"]
  304. dn_1_2 = scenario_2.data_nodes["dn_1"]
  305. dn_2_2 = scenario_2.data_nodes["dn_2"]
  306. dn_3 = scenario_1.data_nodes["dn_3"]
  307. assert dn_3 == scenario_2.data_nodes["dn_3"]
  308. assert dn_1_1.last_edit_date is None
  309. assert dn_2_1.last_edit_date is None
  310. assert dn_1_2.last_edit_date is None
  311. assert dn_2_2.last_edit_date is None
  312. assert dn_3.last_edit_date is None
  313. dn_1_1.last_edit_date = current_datetime + timedelta(1)
  314. dn_2_1.last_edit_date = current_datetime + timedelta(2)
  315. dn_1_2.last_edit_date = current_datetime + timedelta(3)
  316. dn_2_2.last_edit_date = current_datetime + timedelta(4)
  317. dn_3.last_edit_date = current_datetime + timedelta(5)
  318. assert dn_1_1.is_up_to_date
  319. assert dn_2_1.is_up_to_date
  320. assert dn_1_2.is_up_to_date
  321. assert dn_2_2.is_up_to_date
  322. assert dn_3.is_up_to_date
  323. dn_2_1.last_edit_date = current_datetime + timedelta(6)
  324. assert dn_1_1.is_up_to_date
  325. assert dn_2_1.is_up_to_date
  326. assert dn_1_2.is_up_to_date
  327. assert dn_2_2.is_up_to_date
  328. assert not dn_3.is_up_to_date
  329. dn_2_1.last_edit_date = current_datetime + timedelta(2)
  330. dn_2_2.last_edit_date = current_datetime + timedelta(6)
  331. assert dn_1_1.is_up_to_date
  332. assert dn_2_1.is_up_to_date
  333. assert dn_1_2.is_up_to_date
  334. assert dn_2_2.is_up_to_date
  335. assert not dn_3.is_up_to_date
  336. dn_2_2.last_edit_date = current_datetime + timedelta(4)
  337. dn_1_1.last_edit_date = current_datetime + timedelta(6)
  338. assert dn_1_1.is_up_to_date
  339. assert not dn_2_1.is_up_to_date
  340. assert dn_1_2.is_up_to_date
  341. assert dn_2_2.is_up_to_date
  342. assert not dn_3.is_up_to_date
  343. dn_1_2.last_edit_date = current_datetime + timedelta(6)
  344. assert dn_1_1.is_up_to_date
  345. assert not dn_2_1.is_up_to_date
  346. assert dn_1_2.is_up_to_date
  347. assert not dn_2_2.is_up_to_date
  348. assert not dn_3.is_up_to_date
  349. def test_do_not_recompute_data_node_valid_but_continue_sequence_execution(self):
  350. a = Config.configure_data_node("A", "pickle", default_data="A")
  351. b = Config.configure_data_node("B", "pickle")
  352. c = Config.configure_data_node("C", "pickle")
  353. d = Config.configure_data_node("D", "pickle")
  354. task_a_b = Config.configure_task("task_a_b", funct_a_b, input=a, output=b, skippable=True)
  355. task_b_c = Config.configure_task("task_b_c", funct_b_c, input=b, output=c)
  356. task_b_d = Config.configure_task("task_b_d", funct_b_d, input=b, output=d)
  357. scenario_cfg = Config.configure_scenario("scenario", [task_a_b, task_b_c, task_b_d])
  358. scenario = tp.create_scenario(scenario_cfg)
  359. scenario.submit()
  360. assert scenario.A.read() == "A"
  361. assert scenario.B.read() == "B"
  362. assert scenario.C.read() == "C"
  363. assert scenario.D.read() == "D"
  364. scenario.submit()
  365. assert len(tp.get_jobs()) == 6
  366. jobs_and_status = [(job.task.config_id, job.status) for job in tp.get_jobs()]
  367. assert ("task_a_b", tp.Status.COMPLETED) in jobs_and_status
  368. assert ("task_a_b", tp.Status.SKIPPED) in jobs_and_status
  369. assert ("task_b_c", tp.Status.COMPLETED) in jobs_and_status
  370. assert ("task_b_d", tp.Status.COMPLETED) in jobs_and_status
  371. def test_data_node_update_after_writing(self):
  372. dn = FakeDataNode("foo")
  373. _DataManager._set(dn)
  374. assert not _DataManager._get(dn.id).is_ready_for_reading
  375. dn.write("Any data")
  376. assert dn.is_ready_for_reading
  377. assert _DataManager._get(dn.id).is_ready_for_reading
  378. def test_expiration_date_raise_if_never_write(self):
  379. dn = FakeDataNode("foo")
  380. with pytest.raises(NoData):
  381. _ = dn.expiration_date
  382. def test_validity_null_if_never_write(self):
  383. dn = FakeDataNode("foo")
  384. assert dn.validity_period is None
  385. def test_auto_set_and_reload(self, current_datetime):
  386. dn_1 = InMemoryDataNode(
  387. "foo",
  388. scope=Scope.GLOBAL,
  389. id=DataNodeId("an_id"),
  390. owner_id=None,
  391. parent_ids=None,
  392. last_edit_date=current_datetime,
  393. edits=[{"job_id": "a_job_id"}],
  394. edit_in_progress=False,
  395. validity_period=None,
  396. properties={
  397. "name": "foo",
  398. },
  399. )
  400. dm = _DataManager()
  401. dm._set(dn_1)
  402. dn_2 = dm._get(dn_1)
  403. # auto set & reload on scope attribute
  404. assert dn_1.scope == Scope.GLOBAL
  405. assert dn_2.scope == Scope.GLOBAL
  406. dn_1.scope = Scope.CYCLE
  407. assert dn_1.scope == Scope.CYCLE
  408. assert dn_2.scope == Scope.CYCLE
  409. dn_2.scope = Scope.SCENARIO
  410. assert dn_1.scope == Scope.SCENARIO
  411. assert dn_2.scope == Scope.SCENARIO
  412. new_datetime = current_datetime + timedelta(1)
  413. new_datetime_1 = current_datetime + timedelta(3)
  414. # auto set & reload on last_edit_date attribute
  415. assert dn_1.last_edit_date == current_datetime
  416. assert dn_2.last_edit_date == current_datetime
  417. dn_1.last_edit_date = new_datetime_1
  418. assert dn_1.last_edit_date == new_datetime_1
  419. assert dn_2.last_edit_date == new_datetime_1
  420. dn_2.last_edit_date = new_datetime
  421. assert dn_1.last_edit_date == new_datetime
  422. assert dn_2.last_edit_date == new_datetime
  423. # auto set & reload on name attribute
  424. assert dn_1.name == "foo"
  425. assert dn_2.name == "foo"
  426. dn_1.name = "fed"
  427. assert dn_1.name == "fed"
  428. assert dn_2.name == "fed"
  429. dn_2.name = "def"
  430. assert dn_1.name == "def"
  431. assert dn_2.name == "def"
  432. # auto set & reload on parent_ids attribute (set() object does not have auto set yet)
  433. assert dn_1.parent_ids == set()
  434. assert dn_2.parent_ids == set()
  435. dn_1._parent_ids.update(["sc2"])
  436. _DataManager._set(dn_1)
  437. assert dn_1.parent_ids == {"sc2"}
  438. assert dn_2.parent_ids == {"sc2"}
  439. dn_2._parent_ids.clear()
  440. dn_2._parent_ids.update(["sc1"])
  441. _DataManager._set(dn_2)
  442. assert dn_1.parent_ids == {"sc1"}
  443. assert dn_2.parent_ids == {"sc1"}
  444. dn_2._parent_ids.clear()
  445. _DataManager._set(dn_2)
  446. # auto set & reload on edit_in_progress attribute
  447. assert not dn_2.edit_in_progress
  448. assert not dn_1.edit_in_progress
  449. dn_1.edit_in_progress = True
  450. assert dn_1.edit_in_progress
  451. assert dn_2.edit_in_progress
  452. dn_2.unlock_edit()
  453. assert not dn_1.edit_in_progress
  454. assert not dn_2.edit_in_progress
  455. dn_1.lock_edit()
  456. assert dn_1.edit_in_progress
  457. assert dn_2.edit_in_progress
  458. # auto set & reload on validity_period attribute
  459. time_period_1 = timedelta(1)
  460. time_period_2 = timedelta(5)
  461. assert dn_1.validity_period is None
  462. assert dn_2.validity_period is None
  463. dn_1.validity_period = time_period_1
  464. assert dn_1.validity_period == time_period_1
  465. assert dn_2.validity_period == time_period_1
  466. dn_2.validity_period = time_period_2
  467. assert dn_1.validity_period == time_period_2
  468. assert dn_2.validity_period == time_period_2
  469. dn_1.last_edit_date = new_datetime
  470. assert len(dn_1.job_ids) == 1
  471. assert len(dn_2.job_ids) == 1
  472. with dn_1 as dn:
  473. assert dn.config_id == "foo"
  474. assert dn.owner_id is None
  475. assert dn.scope == Scope.SCENARIO
  476. assert dn.last_edit_date == new_datetime
  477. assert dn.name == "def"
  478. assert dn.edit_in_progress
  479. assert dn.validity_period == time_period_2
  480. assert len(dn.job_ids) == 1
  481. assert dn._is_in_context
  482. new_datetime_2 = new_datetime + timedelta(5)
  483. dn.scope = Scope.CYCLE
  484. dn.last_edit_date = new_datetime_2
  485. dn.name = "abc"
  486. dn.edit_in_progress = False
  487. dn.validity_period = None
  488. assert dn.config_id == "foo"
  489. assert dn.owner_id is None
  490. assert dn.scope == Scope.SCENARIO
  491. assert dn.last_edit_date == new_datetime
  492. assert dn.name == "def"
  493. assert dn.edit_in_progress
  494. assert dn.validity_period == time_period_2
  495. assert len(dn.job_ids) == 1
  496. assert dn_1.config_id == "foo"
  497. assert dn_1.owner_id is None
  498. assert dn_1.scope == Scope.CYCLE
  499. assert dn_1.last_edit_date == new_datetime_2
  500. assert dn_1.name == "abc"
  501. assert not dn_1.edit_in_progress
  502. assert dn_1.validity_period is None
  503. assert not dn_1._is_in_context
  504. assert len(dn_1.job_ids) == 1
  505. def test_auto_set_and_reload_properties(self):
  506. dn_1 = InMemoryDataNode("foo", scope=Scope.GLOBAL, properties={"name": "def"})
  507. dm = _DataManager()
  508. dm._set(dn_1)
  509. dn_2 = dm._get(dn_1)
  510. # auto set & reload on properties attribute
  511. assert dn_1.properties == {"name": "def"}
  512. assert dn_2.properties == {"name": "def"}
  513. dn_1._properties["qux"] = 4
  514. assert dn_1.properties["qux"] == 4
  515. assert dn_2.properties["qux"] == 4
  516. assert dn_1.properties == {"qux": 4, "name": "def"}
  517. assert dn_2.properties == {"qux": 4, "name": "def"}
  518. dn_2._properties["qux"] = 5
  519. assert dn_1.properties["qux"] == 5
  520. assert dn_2.properties["qux"] == 5
  521. dn_1.properties["temp_key_1"] = "temp_value_1"
  522. dn_1.properties["temp_key_2"] = "temp_value_2"
  523. assert dn_1.properties == {
  524. "name": "def",
  525. "qux": 5,
  526. "temp_key_1": "temp_value_1",
  527. "temp_key_2": "temp_value_2",
  528. }
  529. assert dn_2.properties == {"name": "def", "qux": 5, "temp_key_1": "temp_value_1", "temp_key_2": "temp_value_2"}
  530. dn_1.properties.pop("temp_key_1")
  531. assert "temp_key_1" not in dn_1.properties.keys()
  532. assert "temp_key_1" not in dn_1.properties.keys()
  533. assert dn_1.properties == {"name": "def", "qux": 5, "temp_key_2": "temp_value_2"}
  534. assert dn_2.properties == {"name": "def", "qux": 5, "temp_key_2": "temp_value_2"}
  535. dn_2.properties.pop("temp_key_2")
  536. assert dn_1.properties == {"qux": 5, "name": "def"}
  537. assert dn_2.properties == {"qux": 5, "name": "def"}
  538. assert "temp_key_2" not in dn_1.properties.keys()
  539. assert "temp_key_2" not in dn_2.properties.keys()
  540. dn_1.properties["temp_key_3"] = 0
  541. assert dn_1.properties == {"qux": 5, "temp_key_3": 0, "name": "def"}
  542. assert dn_2.properties == {"qux": 5, "temp_key_3": 0, "name": "def"}
  543. dn_1.properties.update({"temp_key_3": 1})
  544. assert dn_1.properties == {"qux": 5, "temp_key_3": 1, "name": "def"}
  545. assert dn_2.properties == {"qux": 5, "temp_key_3": 1, "name": "def"}
  546. dn_1.properties.update({})
  547. assert dn_1.properties == {"qux": 5, "temp_key_3": 1, "name": "def"}
  548. assert dn_2.properties == {"qux": 5, "temp_key_3": 1, "name": "def"}
  549. dn_1.properties["temp_key_4"] = 0
  550. dn_1.properties["temp_key_5"] = 0
  551. with dn_1 as dn:
  552. assert dn._is_in_context
  553. assert dn.properties["qux"] == 5
  554. assert dn.properties["temp_key_3"] == 1
  555. assert dn.properties["temp_key_4"] == 0
  556. assert dn.properties["temp_key_5"] == 0
  557. dn.properties["qux"] = 9
  558. dn.properties.pop("temp_key_3")
  559. dn.properties.pop("temp_key_4")
  560. dn.properties.update({"temp_key_4": 1})
  561. dn.properties.update({"temp_key_5": 2})
  562. dn.properties.pop("temp_key_5")
  563. dn.properties.update({})
  564. assert dn.properties["qux"] == 5
  565. assert dn.properties["temp_key_3"] == 1
  566. assert dn.properties["temp_key_4"] == 0
  567. assert dn.properties["temp_key_5"] == 0
  568. assert not dn_1._is_in_context
  569. assert dn_1.properties["qux"] == 9
  570. assert "temp_key_3" not in dn_1.properties.keys()
  571. assert dn_1.properties["temp_key_4"] == 1
  572. assert "temp_key_5" not in dn_1.properties.keys()
  573. def test_get_parents(self, data_node):
  574. with mock.patch("taipy.core.get_parents") as mck:
  575. data_node.get_parents()
  576. mck.assert_called_once_with(data_node)
  577. def test_data_node_with_env_variable_value_not_stored(self):
  578. dn_config = Config.configure_data_node("A", prop="ENV[FOO]")
  579. with mock.patch.dict(os.environ, {"FOO": "bar"}):
  580. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  581. assert dn._properties.data["prop"] == "ENV[FOO]"
  582. assert dn.properties["prop"] == "bar"
  583. def test_path_populated_with_config_default_path(self):
  584. dn_config = Config.configure_data_node("data_node", "pickle", default_path="foo.p")
  585. assert dn_config.default_path == "foo.p"
  586. data_node = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  587. assert data_node.path == "foo.p"
  588. data_node.path = "baz.p"
  589. assert data_node.path == "baz.p"
  590. def test_edit_edit_tracking(self):
  591. dn_config = Config.configure_data_node("A")
  592. data_node = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  593. data_node.write(data="1", job_id="job_1")
  594. data_node.write(data="2", job_id="job_1")
  595. data_node.write(data="3", job_id="job_1")
  596. assert len(data_node.edits) == 3
  597. assert len(data_node.job_ids) == 3
  598. assert data_node.edits[-1] == data_node.get_last_edit()
  599. assert data_node.last_edit_date == data_node.get_last_edit().get("timestamp")
  600. date = datetime(2050, 1, 1, 12, 12)
  601. data_node.write(data="4", timestamp=date, message="This is a comment on this edit", env="staging")
  602. assert len(data_node.edits) == 4
  603. assert len(data_node.job_ids) == 3
  604. assert data_node.edits[-1] == data_node.get_last_edit()
  605. last_edit = data_node.get_last_edit()
  606. assert last_edit["message"] == "This is a comment on this edit"
  607. assert last_edit["env"] == "staging"
  608. assert last_edit["timestamp"] == date
  609. def test_label(self):
  610. a_date = datetime.now()
  611. dn = DataNode(
  612. "foo_bar",
  613. Scope.SCENARIO,
  614. DataNodeId("an_id"),
  615. "a_scenario_id",
  616. {"a_parent_id"},
  617. a_date,
  618. [{"job_id": "a_job_id"}],
  619. edit_in_progress=False,
  620. prop="erty",
  621. name="a name",
  622. )
  623. with mock.patch("taipy.core.get") as get_mck:
  624. class MockOwner:
  625. label = "owner_label"
  626. def get_label(self):
  627. return self.label
  628. get_mck.return_value = MockOwner()
  629. assert dn.get_label() == "owner_label > " + dn.name
  630. assert dn.get_simple_label() == dn.name
  631. def test_explicit_label(self):
  632. a_date = datetime.now()
  633. dn = DataNode(
  634. "foo_bar",
  635. Scope.SCENARIO,
  636. DataNodeId("an_id"),
  637. "a_scenario_id",
  638. {"a_parent_id"},
  639. a_date,
  640. [{"job_id": "a_job_id"}],
  641. edit_in_progress=False,
  642. label="a label",
  643. name="a name",
  644. )
  645. assert dn.get_label() == "a label"
  646. assert dn.get_simple_label() == "a label"
  647. def test_change_data_node_name(self):
  648. cgf = Config.configure_data_node("foo", scope=Scope.GLOBAL)
  649. dn = tp.create_global_data_node(cgf)
  650. dn.name = "bar"
  651. assert dn.name == "bar"
  652. # This new syntax will be the only one allowed: https://github.com/Avaiga/taipy-core/issues/806
  653. dn.properties["name"] = "baz"
  654. assert dn.name == "baz"
  655. def test_locked_data_node_write_should_fail_with_wrong_editor(self):
  656. dn_config = Config.configure_data_node("A")
  657. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  658. dn.lock_edit("editor_1")
  659. # Should raise exception for wrong editor
  660. with pytest.raises(DataNodeIsBeingEdited):
  661. dn.write("data", editor_id="editor_2")
  662. # Should succeed with correct editor
  663. dn.write("data", editor_id="editor_1")
  664. assert dn.read() == "data"
  665. def test_locked_data_node_write_should_fail_before_expiration_date_and_succeed_after(self):
  666. dn_config = Config.configure_data_node("A")
  667. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  668. lock_time = datetime.now()
  669. with freezegun.freeze_time(lock_time):
  670. dn.lock_edit("editor_1")
  671. with freezegun.freeze_time(lock_time + timedelta(minutes=29)):
  672. # Should raise exception for wrong editor and expiration date NOT passed
  673. with pytest.raises(DataNodeIsBeingEdited):
  674. dn.write("data", editor_id="editor_2")
  675. with freezegun.freeze_time(lock_time + timedelta(minutes=31)):
  676. # Should succeed with wrong editor but expiration date passed
  677. dn.write("data", editor_id="editor_2")
  678. assert dn.read() == "data"
  679. def test_locked_data_node_append_should_fail_with_wrong_editor(self):
  680. dn_config = Config.configure_csv_data_node("A")
  681. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  682. first_line = pd.DataFrame(data={'col1': [1], 'col2': [3]})
  683. second_line = pd.DataFrame(data={'col1': [2], 'col2': [4]})
  684. data = pd.DataFrame(data={'col1': [1, 2], 'col2': [3, 4]})
  685. dn.write(first_line)
  686. assert first_line.equals(dn.read())
  687. dn.lock_edit("editor_1")
  688. with pytest.raises(DataNodeIsBeingEdited):
  689. dn.append(second_line, editor_id="editor_2")
  690. dn.append(second_line, editor_id="editor_1")
  691. assert dn.read().equals(data)
  692. def test_locked_data_node_append_should_fail_before_expiration_date_and_succeed_after(self):
  693. dn_config = Config.configure_csv_data_node("A")
  694. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  695. first_line = pd.DataFrame(data={'col1': [1], 'col2': [3]})
  696. second_line = pd.DataFrame(data={'col1': [2], 'col2': [4]})
  697. data = pd.DataFrame(data={'col1': [1, 2], 'col2': [3, 4]})
  698. dn.write(first_line)
  699. assert first_line.equals(dn.read())
  700. lock_time = datetime.now()
  701. with freezegun.freeze_time(lock_time):
  702. dn.lock_edit("editor_1")
  703. with freezegun.freeze_time(lock_time + timedelta(minutes=29)):
  704. # Should raise exception for wrong editor and expiration date NOT passed
  705. with pytest.raises(DataNodeIsBeingEdited):
  706. dn.append(second_line, editor_id="editor_2")
  707. with freezegun.freeze_time(lock_time + timedelta(minutes=31)):
  708. # Should succeed with wrong editor but expiration date passed
  709. dn.append(second_line, editor_id="editor_2")
  710. assert dn.read().equals(data)
  711. def test_orchestrator_write_without_editor_id(self):
  712. dn_config = Config.configure_data_node("A")
  713. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  714. dn.lock_edit("editor_1")
  715. # Orchestrator write without editor_id should succeed
  716. dn.write("orchestrator_data")
  717. assert dn.read() == "orchestrator_data"
  718. def test_editor_fails_writing_a_data_node_locked_by_orchestrator(self):
  719. dn_config = Config.configure_data_node("A")
  720. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  721. dn.lock_edit() # Locked by orchestrator
  722. with pytest.raises(DataNodeIsBeingEdited):
  723. dn.write("data", editor_id="editor_1")
  724. # Orchestrator write without editor_id should succeed
  725. dn.write("orchestrator_data", job_id=JobId("job_1"))
  726. assert dn.read() == "orchestrator_data"
  727. def test_editor_fails_appending_a_data_node_locked_by_orchestrator(self):
  728. dn_config = Config.configure_csv_data_node("A")
  729. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  730. first_line = pd.DataFrame(data={'col1': [1], 'col2': [3]})
  731. second_line = pd.DataFrame(data={'col1': [2], 'col2': [4]})
  732. data = pd.DataFrame(data={'col1': [1, 2], 'col2': [3, 4]})
  733. dn.write(first_line)
  734. assert first_line.equals(dn.read())
  735. dn = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  736. dn.lock_edit() # Locked by orchestrator
  737. with pytest.raises(DataNodeIsBeingEdited):
  738. dn.append(second_line, editor_id="editor_1")
  739. assert dn.read().equals(first_line)
  740. dn.append(second_line, job_id=JobId("job_1"))
  741. assert dn.read().equals(data)
  742. def test_track_edit(self):
  743. dn_config = Config.configure_data_node("A")
  744. data_node = _DataManager._bulk_get_or_create([dn_config])[dn_config]
  745. before = datetime.now()
  746. data_node.track_edit(job_id="job_1")
  747. data_node.track_edit(editor_id="editor_1")
  748. data_node.track_edit(comment="This is a comment on this edit")
  749. data_node.track_edit(editor_id="editor_2", comment="This is another comment on this edit")
  750. data_node.track_edit(editor_id="editor_3", foo="bar")
  751. after = datetime.now()
  752. timestamp = datetime.now()
  753. data_node.track_edit(timestamp=timestamp)
  754. _DataManagerFactory._build_manager()._set(data_node)
  755. # To save the edits because track edit does not save the data node
  756. assert len(data_node.edits) == 6
  757. assert data_node.edits[-1] == data_node.get_last_edit()
  758. assert data_node.last_edit_date == data_node.get_last_edit().get(EDIT_TIMESTAMP_KEY)
  759. edit_0 = data_node.edits[0]
  760. assert len(edit_0) == 2
  761. assert edit_0[EDIT_JOB_ID_KEY] == "job_1"
  762. assert edit_0[EDIT_TIMESTAMP_KEY] >= before
  763. assert edit_0[EDIT_TIMESTAMP_KEY] <= after
  764. edit_1 = data_node.edits[1]
  765. assert len(edit_1) == 2
  766. assert edit_1[EDIT_EDITOR_ID_KEY] == "editor_1"
  767. assert edit_1[EDIT_TIMESTAMP_KEY] >= before
  768. assert edit_1[EDIT_TIMESTAMP_KEY] <= after
  769. edit_2 = data_node.edits[2]
  770. assert len(edit_2) == 2
  771. assert edit_2[EDIT_COMMENT_KEY] == "This is a comment on this edit"
  772. assert edit_2[EDIT_TIMESTAMP_KEY] >= before
  773. assert edit_2[EDIT_TIMESTAMP_KEY] <= after
  774. edit_3 = data_node.edits[3]
  775. assert len(edit_3) == 3
  776. assert edit_3[EDIT_EDITOR_ID_KEY] == "editor_2"
  777. assert edit_3[EDIT_COMMENT_KEY] == "This is another comment on this edit"
  778. assert edit_3[EDIT_TIMESTAMP_KEY] >= before
  779. assert edit_3[EDIT_TIMESTAMP_KEY] <= after
  780. edit_4 = data_node.edits[4]
  781. assert len(edit_4) == 3
  782. assert edit_4[EDIT_EDITOR_ID_KEY] == "editor_3"
  783. assert edit_4["foo"] == "bar"
  784. assert edit_4[EDIT_TIMESTAMP_KEY] >= before
  785. assert edit_4[EDIT_TIMESTAMP_KEY] <= after
  786. edit_5 = data_node.edits[5]
  787. assert len(edit_5) == 1
  788. assert edit_5[EDIT_TIMESTAMP_KEY] == timestamp
  789. def test_normalize_path(self):
  790. dn = DataNode(
  791. config_id="foo_bar",
  792. scope=Scope.SCENARIO,
  793. id=DataNodeId("an_id"),
  794. path=r"data\foo\bar.csv",
  795. )
  796. assert dn.config_id == "foo_bar"
  797. assert dn.scope == Scope.SCENARIO
  798. assert dn.id == "an_id"
  799. assert dn.properties["path"] == "data/foo/bar.csv"