import os import re import json import shutil import requests from bs4 import BeautifulSoup from typing import List rubric = None message = None rubric_file = 'docs/rubric_data.json' discussion_entries_file = 'docs/discussion_entries.json' class DiscussionEntry: def __init__(self, id: int, parent_id: int, name: str, message: str, replies: List): self.id = id self.parent_id = parent_id self.name = name self.message = message self.replies = replies def to_json(self): return { 'id': self.id, 'parent_id': self.parent_id, 'name': self.name, 'message': self.message, 'replies': [reply.to_json() for reply in self.replies] } def dump_json(self, filename): with open(filename, 'w') as f: json.dump(self.to_json(), f) def extract_entries(entries, participants): result = [] for entry in entries: if 'message' in entry and 'deleted' not in entry: id = entry['id'] parent_id = entry['parent_id'] user_id = entry['user_id'] name = next((p['display_name'] for p in participants if p['id'] == user_id), None) message = entry['message'] replies = [] if 'replies' in entry: replies = extract_entries(entry['replies'], participants) result.append(DiscussionEntry(id, parent_id, name, message, replies)) return result def save_messages(entries, group_id=None): for entry in entries: filename = f'docs/{entry.name}.html' if group_id is not None: filename = f'docs/group_{group_id}_{entry.name}.html' with open(filename, 'a+') as f: if entry.parent_id == None: f.write(f'

Student Post: {entry.name}

') f.write(entry.message) f.write('
') else: f.write(f'

Reply to: {entry.parent_id}

') f.write(entry.message) f.write('
') save_messages(entry.replies, group_id) def extract_group_discussions(group_topic_children, headers): group_entries = [] for group_topic in group_topic_children: group_id = group_topic['group_id'] topic_id = group_topic['id'] group_discussion_url = f'{base_url}/api/v1/groups/{group_id}/discussion_topics/{topic_id}/view' group_discussion_response = requests.get(group_discussion_url, headers=headers) if group_discussion_response.ok: group_discussion_data = group_discussion_response.json() entries = extract_entries(group_discussion_data['view'], group_discussion_data['participants']) # Dump JSON data for group-based discussion with open(discussion_entries_file, 'w') as f: json.dump([entry.to_json() for entry in entries], f) group_entries.append({ 'group_id': group_id, 'entries': entries }) return group_entries def extract_individual_discussion(discussion_url, headers): individual_entries = [] discussion_response = requests.get(discussion_url, headers=headers) if discussion_response.ok: discussion_data = discussion_response.json() entries = extract_entries(discussion_data['view'], discussion_data['participants']) # Dump JSON data for individual discussion with open(discussion_entries_file, 'w') as f: json.dump([entry.to_json() for entry in entries], f) individual_entries.extend(entries) return individual_entries def ingest_canvas_discussions(input_url, access_token): global base_url, rubric, message match = re.match(r'https://canvas.illinois.edu/courses/(\d+)/discussion_topics/(\d+)', input_url) if match: course_id, discussion_topic_id = match.groups() else: raise ValueError("Invalid URL") base_url = 'https://canvas.illinois.edu' headers = { 'Authorization': f'Bearer {access_token}' } discussion_url = f'{base_url}/api/v1/courses/{course_id}/discussion_topics/{discussion_topic_id}/view' instruction_url = f'{base_url}/api/v1/courses/{course_id}/discussion_topics/{discussion_topic_id}' instruction_response = requests.get(instruction_url, headers=headers) if instruction_response.ok: instruction_data = instruction_response.json() print(instruction_data) rubric = [] # Extract title if it exists if 'title' in instruction_data: title = instruction_data['title'] rubric = [{'title': title}] if 'description' in instruction_data['assignment']: message_html = instruction_data['assignment']['description'] soup = BeautifulSoup(message_html, 'html.parser') message = soup.get_text() rubric.append({'instruction': message}) if 'rubric' in instruction_data['assignment'] and 'description' in instruction_data['assignment']: rubric.extend(instruction_data['assignment']['rubric']) if 'points_possible' in instruction_data['assignment']: points_possible = instruction_data['assignment']['points_possible'] rubric.append({'points_possible': points_possible}) # Check if the docs folder exists if os.path.exists('docs'): #delete the folder shutil.rmtree('docs') # Create the docs folder os.makedirs('docs') with open(rubric_file, 'w') as f: json.dump(rubric, f) print("Extracted instructions and rubric") else: print(f'Error: {instruction_response.text}') # Check if the discussion is an individual discussion with associated group-based discussions if 'group_topic_children' in instruction_data: # Extract and save group-based discussions group_entries = extract_group_discussions(instruction_data['group_topic_children'], headers) os.makedirs('docs', exist_ok=True) print("Extracted group discussion entries") for group_entry in group_entries: save_messages(group_entry['entries'], group_entry['group_id']) else: # Extract and save standalone individual or group-based discussion individual_entries = extract_individual_discussion(discussion_url, headers) print("Extracted individual discussion entries") os.makedirs('docs', exist_ok=True) save_messages(individual_entries) else: print(f'Error: {instruction_response.text}') def create_vector_store(): return None