def _search(self, term: str): results = self.analyzer.search_similar_content(term) if results: print(f"\nš Search results for 'term':") for result in results: print(f"\n Page result['page_number'] (Similarity: result['similarity_score']:.2f)") print(f" Excerpt: result['excerpt'][:200]...") else: print(f"No results found for 'term'")
def export_to_json(self, output_path: str): """Export all analysis results to JSON file""" output = 'metadata': 'source_file': self.pdf_path, 'total_pages': len(self.pages_text), 'total_words': len(self.full_text.split()) , 'summary': self.create_summary(), 'sections': self.sections, 'key_concepts': self.key_concepts, 'case_studies': self.case_studies, 'study_questions': self.generate_study_questions(), 'full_text_excerpt': self.full_text[:5000] # First 5000 chars with open(output_path, 'w', encoding='utf-8') as f: json.dump(output, f, indent=2, ensure_ascii=False) print(f"Analysis exported to output_path") class UrbanPlanningStudyAssistant: def init (self, analyzer: UrbanPlanningNotesAnalyzer): self.analyzer = analyzer urban planning lecture notes pdf
def _take_quiz(self): questions = self.analyzer.generate_study_questions()[:5] score = 0 print("\nš QUICK QUIZ (5 questions)") print("Answer in your own words, then press Enter for sample answer\n") for i, q in enumerate(questions, 1): print(f"\ni. q['question']") input("Press Enter to see sample answer...") print(f"\n Sample approach: q['hint']") print(" Review the relevant section for complete answer.\n") def main(): # Replace with your PDF path pdf_path = "urban_planning_lecture_notes.pdf" def _search(self, term: str): results = self
def _show_summary(self): summary = self.analyzer.create_summary() print("\nš LECTURE SUMMARY:") print(f" Pages: summary['total_pages']") print(f" Total Words: summary['total_words']:,") print(f" Case Studies: summary['case_studies_count']") print(f"\n Main Topics: ', '.join(summary['key_topics'][:10])") print(f"\n Key Sections: ', '.join(summary['main_sections'][:5])") encoding='utf-8') as f: json.dump(output
def extract_key_concepts(self) -> List[Dict]: """Extract and rank key urban planning concepts""" stop_words = set(stopwords.words('english')) # Urban planning specific terminology planning_terms = [ 'zoning', 'land use', 'transportation', 'infrastructure', 'sustainability', 'urban design', 'smart growth', 'new urbanism', 'gentrification', 'affordable housing', 'public space', 'transit-oriented development', 'mixed-use', 'walkability', 'green infrastructure', 'climate resilience', 'urban renewal', 'community engagement', 'comprehensive plan', 'subdivision', 'environmental impact', 'historic preservation', 'urban sprawl', 'density', 'parking', 'complete streets', 'placemaking' ] # Tokenize and find frequencies words = word_tokenize(self.full_text.lower()) words = [w for w in words if w.isalpha() and w not in stop_words] # Count frequencies of planning terms concept_counts = Counter() for term in planning_terms: count = self.full_text.lower().count(term) if count > 0: concept_counts[term] = count # Extract context for each concept concepts = [] for concept, count in concept_counts.most_common(20): # Find sentences containing the concept sentences = sent_tokenize(self.full_text) context_sentences = [s for s in sentences if concept.lower() in s.lower()] context = context_sentences[:2] if context_sentences else [] concepts.append( 'term': concept, 'frequency': count, 'context': context ) self.key_concepts = concepts return concepts
def _identify_focus_areas(self) -> List[str]: """Identify areas that need more attention based on complexity markers""" complexity_markers = [ 'important', 'crucial', 'essential', 'note that', 'remember', 'key point', 'significant', 'critical', 'fundamental' ] focus_areas = [] sentences = sent_tokenize(self.full_text) for sentence in sentences: for marker in complexity_markers: if marker in sentence.lower(): focus_areas.append(sentence[:100]) break return list(set(focus_areas))[:8]