From 71500dc81c808d6284e6a9dde1b1bec0f54541ed Mon Sep 17 00:00:00 2001 From: Timna Brown <24630902+brown9804@users.noreply.github.com> Date: Mon, 21 Apr 2025 10:18:48 -0600 Subject: [PATCH 1/8] no push needed --- .github/workflows/validate_and_fix_markdown.yml | 3 --- 1 file changed, 3 deletions(-) diff --git a/.github/workflows/validate_and_fix_markdown.yml b/.github/workflows/validate_and_fix_markdown.yml index 9ed1182..3ce58a0 100644 --- a/.github/workflows/validate_and_fix_markdown.yml +++ b/.github/workflows/validate_and_fix_markdown.yml @@ -4,9 +4,6 @@ on: pull_request: branches: - main - push: - branches: - - main permissions: contents: write From 6a3a2892fecaa00e519792a5e0ee207a487f695c Mon Sep 17 00:00:00 2001 From: Timna Brown <24630902+brown9804@users.noreply.github.com> Date: Mon, 21 Apr 2025 10:20:56 -0600 Subject: [PATCH 2/8] no push needed --- .github/workflows/validate_and_fix_notebook.yml | 3 --- 1 file changed, 3 deletions(-) diff --git a/.github/workflows/validate_and_fix_notebook.yml b/.github/workflows/validate_and_fix_notebook.yml index c0e782d..7720e73 100644 --- a/.github/workflows/validate_and_fix_notebook.yml +++ b/.github/workflows/validate_and_fix_notebook.yml @@ -4,9 +4,6 @@ on: pull_request: branches: - main - push: - branches: - - main permissions: contents: write From 0e286bce18c959b0e55ba2db6807b6cd62c25699 Mon Sep 17 00:00:00 2001 From: Timna Brown <24630902+brown9804@users.noreply.github.com> Date: Mon, 21 Apr 2025 10:23:53 -0600 Subject: [PATCH 3/8] format ext --- .github/workflows/validate_and_fix_notebook.yml | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/.github/workflows/validate_and_fix_notebook.yml b/.github/workflows/validate_and_fix_notebook.yml index 7720e73..602d39f 100644 --- a/.github/workflows/validate_and_fix_notebook.yml +++ b/.github/workflows/validate_and_fix_notebook.yml @@ -4,7 +4,7 @@ on: pull_request: branches: - main - + permissions: contents: write @@ -32,12 +32,12 @@ jobs: python -c " import nbformat import glob - for file in glob.glob('**/*.ypyb', recursive=True): + for file in glob.glob('**/*.ipynb', recursive=True): with open(file, 'r') as f: nb = nbformat.read(f, as_version=4) nbformat.validate(nb) - if 'application/vnd.beylor-adapt+notebook' not in nb.metadata: - nb.metadata['application/vnd.beylor-adapt+notebook'] = {'version': '1.0'} + if 'application/vnd.jupyter.widget-state+json' not in nb.metadata: + nb.metadata['application/vnd.jupyter.widget-state+json'] = {'version': '1.0'} with open(file, 'w') as f: nbformat.write(nb, f) " From 654378ef211c6faaffa0a717e32d7c357b43485f Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 21 Apr 2025 16:24:20 +0000 Subject: [PATCH 4/8] Fix notebook format issues --- .../GeneratesRandomData.ipynb | 102 +++++++++++++++++- .../src/fabric-llms-overview_sample.ipynb | 3 + 2 files changed, 104 insertions(+), 1 deletion(-) diff --git a/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb b/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb index 1b00089..8a65895 100644 --- a/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb +++ b/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb @@ -1 +1,101 @@ -{"cells":[{"cell_type":"code","source":["# Generates Dummy json file in Files/\n","\n","# Import necessary libraries\n","from pyspark.sql import SparkSession\n","from pyspark.sql.types import *\n","import random\n","from datetime import datetime, timedelta\n","\n","# Initialize Spark session (if not already initialized)\n","spark = SparkSession.builder.appName(\"GenerateRandomData\").getOrCreate()\n","\n","# Function to generate random data\n","def generate_random_data(num_entries):\n"," data = []\n"," for i in range(1, num_entries + 1):\n"," name = f\"User{i}\"\n"," entry = {\n"," \"id\": i,\n"," \"name\": name,\n"," \"age\": random.randint(18, 65),\n"," \"email\": f\"{name.lower()}@example.com\",\n"," \"created_at\": (datetime.now() - timedelta(days=random.randint(0, 365))).strftime(\"%Y-%m-%d %H:%M:%S\")\n"," }\n"," data.append(entry)\n"," return data\n","\n","# Generate 10 random entries\n","random_data = generate_random_data(10)\n","\n","# Define schema for the DataFrame\n","schema = StructType([\n"," StructField(\"id\", IntegerType(), True),\n"," StructField(\"name\", StringType(), True),\n"," StructField(\"age\", IntegerType(), True),\n"," StructField(\"email\", StringType(), True),\n"," StructField(\"created_at\", StringType(), True)\n","])\n","\n","# Create a DataFrame from the random data\n","df_random_data = spark.createDataFrame(random_data, schema=schema)\n","\n","# Write the DataFrame to the Lakehouse in the specified path\n","output_path = \"abfss://{WORKSPACE-NAME}@onelake.dfs.fabric.microsoft.com/raw_Bronze.Lakehouse/Files/random_data\" # Replace {WORKSPACE-NAME}\n","df_random_data.write.format(\"delta\").mode(\"overwrite\").save(output_path)\n","\n","print(f\"Random data has been saved to the Lakehouse at '{output_path}'.\")"],"outputs":[],"execution_count":null,"metadata":{"microsoft":{"language":"python","language_group":"synapse_pyspark"}},"id":"8d820f25-3c2e-45b3-8a08-af78f0d45e1d"}],"metadata":{"kernel_info":{"name":"synapse_pyspark"},"kernelspec":{"name":"synapse_pyspark","language":"Python","display_name":"Synapse PySpark"},"language_info":{"name":"python"},"microsoft":{"language":"python","language_group":"synapse_pyspark","ms_spell_check":{"ms_spell_check_language":"en"}},"nteract":{"version":"nteract-front-end@1.0.0"},"spark_compute":{"compute_id":"/trident/default","session_options":{"conf":{"spark.synapse.nbs.session.timeout":"1200000"}}},"dependencies":{}},"nbformat":4,"nbformat_minor":5} +{ + "cells": [ + { + "cell_type": "code", + "execution_count": null, + "id": "8d820f25-3c2e-45b3-8a08-af78f0d45e1d", + "metadata": { + "microsoft": { + "language": "python", + "language_group": "synapse_pyspark" + } + }, + "outputs": [], + "source": [ + "# Generates Dummy json file in Files/\n", + "\n", + "# Import necessary libraries\n", + "from pyspark.sql import SparkSession\n", + "from pyspark.sql.types import *\n", + "import random\n", + "from datetime import datetime, timedelta\n", + "\n", + "# Initialize Spark session (if not already initialized)\n", + "spark = SparkSession.builder.appName(\"GenerateRandomData\").getOrCreate()\n", + "\n", + "# Function to generate random data\n", + "def generate_random_data(num_entries):\n", + " data = []\n", + " for i in range(1, num_entries + 1):\n", + " name = f\"User{i}\"\n", + " entry = {\n", + " \"id\": i,\n", + " \"name\": name,\n", + " \"age\": random.randint(18, 65),\n", + " \"email\": f\"{name.lower()}@example.com\",\n", + " \"created_at\": (datetime.now() - timedelta(days=random.randint(0, 365))).strftime(\"%Y-%m-%d %H:%M:%S\")\n", + " }\n", + " data.append(entry)\n", + " return data\n", + "\n", + "# Generate 10 random entries\n", + "random_data = generate_random_data(10)\n", + "\n", + "# Define schema for the DataFrame\n", + "schema = StructType([\n", + " StructField(\"id\", IntegerType(), True),\n", + " StructField(\"name\", StringType(), True),\n", + " StructField(\"age\", IntegerType(), True),\n", + " StructField(\"email\", StringType(), True),\n", + " StructField(\"created_at\", StringType(), True)\n", + "])\n", + "\n", + "# Create a DataFrame from the random data\n", + "df_random_data = spark.createDataFrame(random_data, schema=schema)\n", + "\n", + "# Write the DataFrame to the Lakehouse in the specified path\n", + "output_path = \"abfss://{WORKSPACE-NAME}@onelake.dfs.fabric.microsoft.com/raw_Bronze.Lakehouse/Files/random_data\" # Replace {WORKSPACE-NAME}\n", + "df_random_data.write.format(\"delta\").mode(\"overwrite\").save(output_path)\n", + "\n", + "print(f\"Random data has been saved to the Lakehouse at '{output_path}'.\")" + ] + } + ], + "metadata": { + "application/vnd.jupyter.widget-state+json": { + "version": "1.0" + }, + "dependencies": {}, + "kernel_info": { + "name": "synapse_pyspark" + }, + "kernelspec": { + "display_name": "Synapse PySpark", + "language": "Python", + "name": "synapse_pyspark" + }, + "language_info": { + "name": "python" + }, + "microsoft": { + "language": "python", + "language_group": "synapse_pyspark", + "ms_spell_check": { + "ms_spell_check_language": "en" + } + }, + "nteract": { + "version": "nteract-front-end@1.0.0" + }, + "spark_compute": { + "compute_id": "/trident/default", + "session_options": { + "conf": { + "spark.synapse.nbs.session.timeout": "1200000" + } + } + } + }, + "nbformat": 4, + "nbformat_minor": 5 +} diff --git a/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb b/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb index 7b0c18d..b97781f 100644 --- a/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb +++ b/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb @@ -1152,6 +1152,9 @@ } ], "metadata": { + "application/vnd.jupyter.widget-state+json": { + "version": "1.0" + }, "dependencies": { "environment": { "environmentId": "766562be-9e21-456c-b270-cac7e4bf8d18", From 9b6552e5468ffc6049588ab911c9f37340396ba0 Mon Sep 17 00:00:00 2001 From: Timna Brown <24630902+brown9804@users.noreply.github.com> Date: Mon, 21 Apr 2025 10:28:17 -0600 Subject: [PATCH 5/8] testing strig --- .github/workflows/validate_and_fix_notebook.yml | 6 +++--- 1 file changed, 3 insertions(+), 3 deletions(-) diff --git a/.github/workflows/validate_and_fix_notebook.yml b/.github/workflows/validate_and_fix_notebook.yml index 602d39f..dad3162 100644 --- a/.github/workflows/validate_and_fix_notebook.yml +++ b/.github/workflows/validate_and_fix_notebook.yml @@ -4,7 +4,7 @@ on: pull_request: branches: - main - + permissions: contents: write @@ -36,8 +36,8 @@ jobs: with open(file, 'r') as f: nb = nbformat.read(f, as_version=4) nbformat.validate(nb) - if 'application/vnd.jupyter.widget-state+json' not in nb.metadata: - nb.metadata['application/vnd.jupyter.widget-state+json'] = {'version': '1.0'} + if 'application/vnd.jupyter.widget-state+json' not in nb.metadata.get('widgets', {}): + nb.metadata['widgets'] = {'application/vnd.jupyter.widget-state+json': {'version': '1.0'}} with open(file, 'w') as f: nbformat.write(nb, f) " From 2e79650f152d499dfd99373bce9a5dd9d0e3ead4 Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 21 Apr 2025 16:28:44 +0000 Subject: [PATCH 6/8] Fix notebook format issues --- .../FabricActivatorRulePipeline/GeneratesRandomData.ipynb | 5 +++++ .../AI_integration/src/fabric-llms-overview_sample.ipynb | 6 +++++- 2 files changed, 10 insertions(+), 1 deletion(-) diff --git a/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb b/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb index 8a65895..4dcd669 100644 --- a/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb +++ b/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb @@ -94,6 +94,11 @@ "spark.synapse.nbs.session.timeout": "1200000" } } + }, + "widgets": { + "application/vnd.jupyter.widget-state+json": { + "version": "1.0" + } } }, "nbformat": 4, diff --git a/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb b/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb index b97781f..fce83f2 100644 --- a/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb +++ b/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb @@ -1190,7 +1190,11 @@ } } }, - "widgets": {} + "widgets": { + "application/vnd.jupyter.widget-state+json": { + "version": "1.0" + } + } }, "nbformat": 4, "nbformat_minor": 5 From f7cf43a82bef60126d8b321b6cecec38b34dfbcc Mon Sep 17 00:00:00 2001 From: Timna Brown <24630902+brown9804@users.noreply.github.com> Date: Mon, 21 Apr 2025 10:32:28 -0600 Subject: [PATCH 7/8] adding metadata --- .github/workflows/validate_and_fix_notebook.yml | 7 +++++-- 1 file changed, 5 insertions(+), 2 deletions(-) diff --git a/.github/workflows/validate_and_fix_notebook.yml b/.github/workflows/validate_and_fix_notebook.yml index dad3162..9bf6902 100644 --- a/.github/workflows/validate_and_fix_notebook.yml +++ b/.github/workflows/validate_and_fix_notebook.yml @@ -36,8 +36,11 @@ jobs: with open(file, 'r') as f: nb = nbformat.read(f, as_version=4) nbformat.validate(nb) - if 'application/vnd.jupyter.widget-state+json' not in nb.metadata.get('widgets', {}): - nb.metadata['widgets'] = {'application/vnd.jupyter.widget-state+json': {'version': '1.0'}} + if 'widgets' in nb.metadata: + if 'application/vnd.jupyter.widget-state+json' not in nb.metadata['widgets']: + nb.metadata['widgets']['application/vnd.jupyter.widget-state+json'] = {'version': '1.0', 'state': {}} + elif 'state' not in nb.metadata['widgets']['application/vnd.jupyter.widget-state+json']: + nb.metadata['widgets']['application/vnd.jupyter.widget-state+json']['state'] = {} with open(file, 'w') as f: nbformat.write(nb, f) " From 681c1d03024afb829212d3ef69e6f070a68bf07f Mon Sep 17 00:00:00 2001 From: "github-actions[bot]" Date: Mon, 21 Apr 2025 16:32:50 +0000 Subject: [PATCH 8/8] Fix notebook format issues --- .../FabricActivatorRulePipeline/GeneratesRandomData.ipynb | 1 + .../AI_integration/src/fabric-llms-overview_sample.ipynb | 1 + 2 files changed, 2 insertions(+) diff --git a/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb b/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb index 4dcd669..6cc6a2c 100644 --- a/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb +++ b/Monitoring-Observability/FabricActivatorRulePipeline/GeneratesRandomData.ipynb @@ -97,6 +97,7 @@ }, "widgets": { "application/vnd.jupyter.widget-state+json": { + "state": {}, "version": "1.0" } } diff --git a/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb b/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb index fce83f2..8146c08 100644 --- a/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb +++ b/Workloads-Specific/DataScience/AI_integration/src/fabric-llms-overview_sample.ipynb @@ -1192,6 +1192,7 @@ }, "widgets": { "application/vnd.jupyter.widget-state+json": { + "state": {}, "version": "1.0" } }