conftest.py 15 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451
  1. # Copyright 2023 Avaiga Private Limited
  2. #
  3. # Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
  4. # the License. You may obtain a copy of the License at
  5. #
  6. # http://www.apache.org/licenses/LICENSE-2.0
  7. #
  8. # Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on
  9. # an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the
  10. # specific language governing permissions and limitations under the License.
  11. import os
  12. import pickle
  13. import shutil
  14. from datetime import datetime
  15. from queue import Queue
  16. import pandas as pd
  17. import pytest
  18. from sqlalchemy import create_engine, text
  19. from src.taipy.core._core import Core
  20. from src.taipy.core._orchestrator._orchestrator_factory import _OrchestratorFactory
  21. from src.taipy.core._repository.db._sql_connection import _SQLConnection
  22. from src.taipy.core._version._version import _Version
  23. from src.taipy.core._version._version_manager_factory import _VersionManagerFactory
  24. from src.taipy.core.config import (
  25. CoreSection,
  26. DataNodeConfig,
  27. JobConfig,
  28. MigrationConfig,
  29. ScenarioConfig,
  30. TaskConfig,
  31. _ConfigIdChecker,
  32. _CoreSectionChecker,
  33. _DataNodeConfigChecker,
  34. _JobConfigChecker,
  35. _ScenarioConfigChecker,
  36. _TaskConfigChecker,
  37. )
  38. from src.taipy.core.cycle._cycle_manager_factory import _CycleManagerFactory
  39. from src.taipy.core.cycle._cycle_model import _CycleModel
  40. from src.taipy.core.cycle.cycle import Cycle
  41. from src.taipy.core.cycle.cycle_id import CycleId
  42. from src.taipy.core.data._data_manager_factory import _DataManagerFactory
  43. from src.taipy.core.data._data_model import _DataNodeModel
  44. from src.taipy.core.data.in_memory import InMemoryDataNode
  45. from src.taipy.core.job._job_manager_factory import _JobManagerFactory
  46. from src.taipy.core.job.job import Job
  47. from src.taipy.core.job.job_id import JobId
  48. from src.taipy.core.notification.notifier import Notifier
  49. from src.taipy.core.scenario._scenario_manager_factory import _ScenarioManagerFactory
  50. from src.taipy.core.scenario._scenario_model import _ScenarioModel
  51. from src.taipy.core.scenario.scenario import Scenario
  52. from src.taipy.core.scenario.scenario_id import ScenarioId
  53. from src.taipy.core.sequence._sequence_manager_factory import _SequenceManagerFactory
  54. from src.taipy.core.sequence.sequence import Sequence
  55. from src.taipy.core.sequence.sequence_id import SequenceId
  56. from src.taipy.core.submission._submission_manager_factory import _SubmissionManagerFactory
  57. from src.taipy.core.submission._submission_model import _SubmissionModel
  58. from src.taipy.core.task._task_manager_factory import _TaskManagerFactory
  59. from src.taipy.core.task.task import Task
  60. from taipy.config import _inject_section
  61. from taipy.config._config import _Config
  62. from taipy.config._serializer._toml_serializer import _TomlSerializer
  63. from taipy.config.checker._checker import _Checker
  64. from taipy.config.checker.issue_collector import IssueCollector
  65. from taipy.config.common.frequency import Frequency
  66. from taipy.config.common.scope import Scope
  67. from taipy.config.config import Config
  68. current_time = datetime.now()
  69. _OrchestratorFactory._build_orchestrator()
  70. @pytest.fixture(scope="function")
  71. def csv_file(tmpdir_factory) -> str:
  72. csv = pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}])
  73. fn = tmpdir_factory.mktemp("data").join("df.csv")
  74. csv.to_csv(str(fn), index=False)
  75. return fn.strpath
  76. @pytest.fixture(scope="function")
  77. def excel_file(tmpdir_factory) -> str:
  78. excel = pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}])
  79. fn = tmpdir_factory.mktemp("data").join("df.xlsx")
  80. excel.to_excel(str(fn), index=False)
  81. return fn.strpath
  82. @pytest.fixture(scope="function")
  83. def excel_file_with_sheet_name(tmpdir_factory) -> str:
  84. excel = pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}])
  85. fn = tmpdir_factory.mktemp("data").join("df.xlsx")
  86. excel.to_excel(str(fn), sheet_name="sheet_name", index=False)
  87. return fn.strpath
  88. @pytest.fixture(scope="function")
  89. def json_file(tmpdir_factory) -> str:
  90. json_data = pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}])
  91. fn = tmpdir_factory.mktemp("data").join("df.json")
  92. json_data.to_json(str(fn), orient="records")
  93. return fn.strpath
  94. @pytest.fixture(scope="function")
  95. def excel_file_with_multi_sheet(tmpdir_factory) -> str:
  96. excel_multi_sheet = {
  97. "Sheet1": pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}]),
  98. "Sheet2": pd.DataFrame([{"a": 7, "b": 8, "c": 9}, {"a": 10, "b": 11, "c": 12}]),
  99. }
  100. fn = tmpdir_factory.mktemp("data").join("df.xlsx")
  101. with pd.ExcelWriter(str(fn)) as writer:
  102. for key in excel_multi_sheet.keys():
  103. excel_multi_sheet[key].to_excel(writer, key, index=False)
  104. return fn.strpath
  105. @pytest.fixture(scope="function")
  106. def pickle_file_path(tmpdir_factory) -> str:
  107. data = pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}])
  108. fn = tmpdir_factory.mktemp("data").join("df.p")
  109. with open(str(fn), "wb") as f:
  110. pickle.dump(data, f)
  111. return fn.strpath
  112. @pytest.fixture(scope="function")
  113. def parquet_file_path(tmpdir_factory) -> str:
  114. data = pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}])
  115. fn = tmpdir_factory.mktemp("data").join("df.parquet")
  116. data.to_parquet(str(fn))
  117. return fn.strpath
  118. @pytest.fixture(scope="function")
  119. def tmp_sqlite_db_file_path(tmpdir_factory):
  120. fn = tmpdir_factory.mktemp("data")
  121. db_name = "df"
  122. file_extension = ".db"
  123. db = create_engine("sqlite:///" + os.path.join(fn.strpath, f"{db_name}{file_extension}"))
  124. conn = db.connect()
  125. conn.execute(text("CREATE TABLE example (foo int, bar int);"))
  126. conn.execute(text("INSERT INTO example (foo, bar) VALUES (1, 2);"))
  127. conn.execute(text("INSERT INTO example (foo, bar) VALUES (3, 4);"))
  128. conn.commit()
  129. conn.close()
  130. db.dispose()
  131. return fn.strpath, db_name, file_extension
  132. @pytest.fixture(scope="function")
  133. def tmp_sqlite_sqlite3_file_path(tmpdir_factory):
  134. fn = tmpdir_factory.mktemp("data")
  135. db_name = "df"
  136. file_extension = ".sqlite3"
  137. db = create_engine("sqlite:///" + os.path.join(fn.strpath, f"{db_name}{file_extension}"))
  138. conn = db.connect()
  139. conn.execute(text("CREATE TABLE example (foo int, bar int);"))
  140. conn.execute(text("INSERT INTO example (foo, bar) VALUES (1, 2);"))
  141. conn.execute(text("INSERT INTO example (foo, bar) VALUES (3, 4);"))
  142. conn.commit()
  143. conn.close()
  144. db.dispose()
  145. return fn.strpath, db_name, file_extension
  146. @pytest.fixture(scope="function")
  147. def default_data_frame():
  148. return pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}])
  149. @pytest.fixture(scope="function")
  150. def default_multi_sheet_data_frame():
  151. return {
  152. "Sheet1": pd.DataFrame([{"a": 1, "b": 2, "c": 3}, {"a": 4, "b": 5, "c": 6}]),
  153. "Sheet2": pd.DataFrame([{"a": 7, "b": 8, "c": 9}, {"a": 10, "b": 11, "c": 12}]),
  154. }
  155. @pytest.fixture(scope="session", autouse=True)
  156. def cleanup_files():
  157. yield
  158. if os.path.exists(".data"):
  159. shutil.rmtree(".data", ignore_errors=True)
  160. if os.path.exists(".my_data"):
  161. shutil.rmtree(".my_data", ignore_errors=True)
  162. @pytest.fixture(scope="function")
  163. def current_datetime():
  164. return current_time
  165. @pytest.fixture(scope="function")
  166. def scenario(cycle):
  167. return Scenario(
  168. "sc",
  169. set(),
  170. {},
  171. set(),
  172. ScenarioId("sc_id"),
  173. current_time,
  174. is_primary=False,
  175. tags={"foo"},
  176. version="random_version_number",
  177. cycle=None,
  178. )
  179. @pytest.fixture(scope="function")
  180. def data_node():
  181. return InMemoryDataNode("data_node_config_id", Scope.SCENARIO, version="random_version_number")
  182. @pytest.fixture(scope="function")
  183. def data_node_model():
  184. return _DataNodeModel(
  185. "my_dn_id",
  186. "test_data_node",
  187. Scope.SCENARIO,
  188. "csv",
  189. "name",
  190. "owner_id",
  191. list({"parent_id_1", "parent_id_2"}),
  192. datetime(1985, 10, 14, 2, 30, 0).isoformat(),
  193. [dict(timestamp=datetime(1985, 10, 14, 2, 30, 0).isoformat(), job_id="job_id")],
  194. "latest",
  195. None,
  196. None,
  197. False,
  198. {"path": "/path", "has_header": True, "prop": "ENV[FOO]", "exposed_type": "pandas"},
  199. )
  200. @pytest.fixture(scope="function")
  201. def task(data_node):
  202. dn = InMemoryDataNode("dn_config_id", Scope.SCENARIO, version="random_version_number")
  203. return Task("task_config_id", {}, print, [data_node], [dn])
  204. @pytest.fixture(scope="function")
  205. def scenario_model(cycle):
  206. return _ScenarioModel(
  207. ScenarioId("sc_id"),
  208. "sc",
  209. set(),
  210. set(),
  211. {},
  212. creation_date=current_time.isoformat(),
  213. primary_scenario=False,
  214. subscribers=[],
  215. tags=["foo"],
  216. version="random_version_number",
  217. cycle=None,
  218. )
  219. @pytest.fixture(scope="function")
  220. def cycle():
  221. example_date = datetime.fromisoformat("2021-11-11T11:11:01.000001")
  222. return Cycle(
  223. Frequency.DAILY,
  224. {},
  225. creation_date=example_date,
  226. start_date=example_date,
  227. end_date=example_date,
  228. name="cc",
  229. id=CycleId("cc_id"),
  230. )
  231. @pytest.fixture(scope="class")
  232. def sequence():
  233. return Sequence(
  234. {},
  235. [],
  236. SequenceId("sequence_id"),
  237. owner_id="owner_id",
  238. parent_ids=set(["parent_id_1", "parent_id_2"]),
  239. version="random_version_number",
  240. )
  241. @pytest.fixture(scope="function")
  242. def job(task):
  243. return Job(JobId("job"), task, "foo", "bar", version="random_version_number")
  244. @pytest.fixture(scope="function")
  245. def _version():
  246. return _Version(id="foo", config=Config._applied_config)
  247. @pytest.fixture(scope="function")
  248. def cycle_model():
  249. return _CycleModel(
  250. CycleId("cc_id"),
  251. "cc",
  252. Frequency.DAILY,
  253. {},
  254. creation_date="2021-11-11T11:11:01.000001",
  255. start_date="2021-11-11T11:11:01.000001",
  256. end_date="2021-11-11T11:11:01.000001",
  257. )
  258. @pytest.fixture(scope="function")
  259. def tmp_sqlite(tmpdir_factory):
  260. fn = tmpdir_factory.mktemp("db")
  261. return os.path.join(fn.strpath, "test.db")
  262. @pytest.fixture(scope="function", autouse=True)
  263. def clean_repository():
  264. from sqlalchemy.orm import close_all_sessions
  265. close_all_sessions()
  266. init_config()
  267. init_orchestrator()
  268. init_managers()
  269. init_config()
  270. init_notifier()
  271. yield
  272. def init_config():
  273. Config.unblock_update()
  274. Config._default_config = _Config()._default_config()
  275. Config._python_config = _Config()
  276. Config._file_config = _Config()
  277. Config._env_file_config = _Config()
  278. Config._applied_config = _Config()
  279. Config._collector = IssueCollector()
  280. Config._serializer = _TomlSerializer()
  281. _Checker._checkers = []
  282. _inject_section(
  283. JobConfig, "job_config", JobConfig("development"), [("configure_job_executions", JobConfig._configure)], True
  284. )
  285. _inject_section(
  286. CoreSection,
  287. "core",
  288. CoreSection.default_config(),
  289. [("configure_core", CoreSection._configure)],
  290. add_to_unconflicted_sections=True,
  291. )
  292. _inject_section(
  293. DataNodeConfig,
  294. "data_nodes",
  295. DataNodeConfig.default_config(),
  296. [
  297. ("configure_data_node", DataNodeConfig._configure),
  298. ("configure_data_node_from", DataNodeConfig._configure_from),
  299. ("set_default_data_node_configuration", DataNodeConfig._set_default_configuration),
  300. ("configure_csv_data_node", DataNodeConfig._configure_csv),
  301. ("configure_json_data_node", DataNodeConfig._configure_json),
  302. ("configure_sql_table_data_node", DataNodeConfig._configure_sql_table),
  303. ("configure_sql_data_node", DataNodeConfig._configure_sql),
  304. ("configure_mongo_collection_data_node", DataNodeConfig._configure_mongo_collection),
  305. ("configure_in_memory_data_node", DataNodeConfig._configure_in_memory),
  306. ("configure_pickle_data_node", DataNodeConfig._configure_pickle),
  307. ("configure_excel_data_node", DataNodeConfig._configure_excel),
  308. ("configure_generic_data_node", DataNodeConfig._configure_generic),
  309. ],
  310. )
  311. _inject_section(
  312. TaskConfig,
  313. "tasks",
  314. TaskConfig.default_config(),
  315. [
  316. ("configure_task", TaskConfig._configure),
  317. ("set_default_task_configuration", TaskConfig._set_default_configuration),
  318. ],
  319. )
  320. _inject_section(
  321. ScenarioConfig,
  322. "scenarios",
  323. ScenarioConfig.default_config(),
  324. [
  325. ("configure_scenario", ScenarioConfig._configure),
  326. ("set_default_scenario_configuration", ScenarioConfig._set_default_configuration),
  327. ],
  328. )
  329. _inject_section(
  330. MigrationConfig,
  331. "migration_functions",
  332. MigrationConfig.default_config(),
  333. [("add_migration_function", MigrationConfig._add_migration_function)],
  334. True,
  335. )
  336. _Checker.add_checker(_ConfigIdChecker)
  337. _Checker.add_checker(_CoreSectionChecker)
  338. _Checker.add_checker(_DataNodeConfigChecker)
  339. _Checker.add_checker(_JobConfigChecker)
  340. # We don't need to add _MigrationConfigChecker because it is run only when the Core service is run.
  341. _Checker.add_checker(_TaskConfigChecker)
  342. _Checker.add_checker(_ScenarioConfigChecker)
  343. Config.configure_core(read_entity_retry=0)
  344. Core._is_running = False
  345. def init_managers():
  346. _CycleManagerFactory._build_manager()._delete_all()
  347. _ScenarioManagerFactory._build_manager()._delete_all()
  348. _SequenceManagerFactory._build_manager()._delete_all()
  349. _JobManagerFactory._build_manager()._delete_all()
  350. _TaskManagerFactory._build_manager()._delete_all()
  351. _DataManagerFactory._build_manager()._delete_all()
  352. _VersionManagerFactory._build_manager()._delete_all()
  353. _SubmissionManagerFactory._build_manager()._delete_all()
  354. def init_orchestrator():
  355. if _OrchestratorFactory._orchestrator is None:
  356. _OrchestratorFactory._build_orchestrator()
  357. _OrchestratorFactory._build_dispatcher()
  358. _OrchestratorFactory._orchestrator.jobs_to_run = Queue()
  359. _OrchestratorFactory._orchestrator.blocked_jobs = []
  360. def init_notifier():
  361. Notifier._topics_registrations_list = {}
  362. @pytest.fixture
  363. def sql_engine():
  364. return create_engine("sqlite:///:memory:")
  365. @pytest.fixture
  366. def init_sql_repo(tmp_sqlite):
  367. Config.configure_core(repository_type="sql", repository_properties={"db_location": tmp_sqlite})
  368. # Clean SQLite database
  369. if _SQLConnection._connection:
  370. _SQLConnection._connection.close()
  371. _SQLConnection._connection = None
  372. _SQLConnection.init_db()
  373. return tmp_sqlite