scaffold_ai
Scaffold AI: Curriculum Recommendation Tool for Sustainability and Climate Resilience
Science Score: 44.0%
This score indicates how likely this project is to be science-related based on various indicators:
-
✓CITATION.cff file
Found CITATION.cff file -
✓codemeta.json file
Found codemeta.json file -
✓.zenodo.json file
Found .zenodo.json file -
○DOI references
-
○Academic publication links
-
○Academic email domains
-
○Institutional organization owner
-
○JOSS paper metadata
-
○Scientific vocabulary similarity
Low similarity (10.4%) to scientific vocabulary
Keywords
Repository
Scaffold AI: Curriculum Recommendation Tool for Sustainability and Climate Resilience
Basic Info
- Host: GitHub
- Owner: kevinmastascusa
- Language: Python
- Default Branch: main
- Homepage: https://research.coe.drexel.edu/caee/circlelab/
- Size: 15.6 MB
Statistics
- Stars: 1
- Watchers: 0
- Forks: 0
- Open Issues: 8
- Releases: 3
Topics
Metadata Files
README.md
# 🌱 Scaffold AI: Curriculum Recommendation Tool for Sustainability and Climate Resilience
[](https://www.python.org/downloads/)
[](https://flask.palletsprojects.com/)
[](https://huggingface.co/TinyLlama/TinyLlama-1.1B-Chat-v1.0)
[](https://github.com/facebookresearch/faiss)
[](https://github.com/kevinmastascusa/scaffold_ai)
[](https://opensource.org/licenses/MIT)
[](https://github.com/kevinmastascusa/scaffold_ai)
[](https://github.com/kevinmastascusa/scaffold_ai)
[](https://github.com/kevinmastascusa/scaffold_ai/issues)
[](https://github.com/kevinmastascusa/scaffold_ai/pulls)
**Collaborators:** Kevin Mastascusa, Joseph Di Stefano
**Date:** 6/26/2025 | **Last Updated:** 7/20/2025
📋 Table of Contents
- 🌍 Project Overview
- 🎉 Current Status
- ✨ Key Features & Recent Improvements
- 🚀 Getting Started
- 🎯 Goals and Objectives
- 🏗️ System Architecture
- 🤖 Large Language Model (LLM)
- 🔗 Citation Tracking and Transparency
- 🔄 Technical Workflow
- 📅 Project Timeline
- 📈 Evaluation Overview
- ✅ Expected Outcomes
- 🧾 TODO Section
- 🔍 Project Validation
- 📝 Model Version/Hash Logging
- ⚡ Model Benchmarking
🌍 Project Overview
This project involves developing a specialized large language model (LLM)-based tool to assist educators in integrating sustainability and climate resilience topics into academic programs. The tool leverages state-of-the-art AI techniques to recommend high-quality, literature-backed educational materials, case studies, and project ideas.
🎉 Current Status: FULLY OPERATIONAL WITH LLM INTEGRATION
✅ What's Working Now:
- 🤖 LLM-Powered Responses: Intelligent, contextual responses generated by TinyLlama model
- 🎨 Modern Chat Interface: Beautiful, responsive web interface with Scaffold AI branding
- 🔍 Advanced Search System: Vector-based search through sustainability research database
- 📚 Complete Source Attribution: Detailed source information with authors, titles, DOIs, and citations
- 💬 Real-time Chat: Smooth conversation flow with proper error handling
- 📊 Research-Based Responses: All responses backed by academic sustainability research
- 🔄 Robust Fallback System: Graceful degradation if LLM fails
- 🔧 Debug & Reset Features: Easy troubleshooting and conversation management
🚀 Quick Start (Production Ready):
```bash
Start the enhanced UI
cd frontend export HUGGINGFACETOKEN=yourtokenhere python appenhanced_simple.py --port 5003 ```
Access the UI: http://localhost:5003 (or your preferred port)
🎬 Quick Demo
**Experience the power of AI-driven sustainability education recommendations!**
🚀 Try It Now:
- Start the server using the Quick Start commands above
- Open your browser to
http://localhost:5003 - Ask questions like:
- "How can I incorporate sustainability in my Fluid Mechanics course?"
- "What are some climate education strategies for engineering students?"
- "Show me examples of green technology integration in thermodynamics"
- Get research-backed responses with complete source citations
💡 Example Interaction:
``` User: "How can I teach sustainability in mechanical engineering?"
Scaffold AI: "Based on research from the ASEE database, here are proven strategies:
Core Course Integration: Incorporate sustainability into thermodynamics, fluid mechanics, and materials science courses...
Project-Based Learning: Design assignments around renewable energy, green manufacturing, and life cycle analysis...
[Source: "Sustainability incorporation in courses in mechanical civil and environmental engineering insights from AASHE STARS data" - ASEE Publications] ```
✨ Key Features & Recent Improvements
🎛️ Model Control Panel (NEW!)
- Dynamic Model Selection: Switch between different LLM and embedding models
- Real-time Parameter Tuning: Adjust temperature, max tokens, chunk size, and overlap
- Visual Status Indicators: See model status (Active/Updating/Error) at a glance
- One-click Apply: Instantly apply model changes with API integration
- Settings Persistence: All settings saved to localStorage for consistency
- Reset to Defaults: Quick reset button for parameter restoration
🔔 Enhanced Notification System
- Animated Notifications: Smooth slide-in/slide-out animations
- Multiple Types: Success, Error, and Info notifications with appropriate styling
- Auto-dismiss: Notifications disappear after 5 seconds
- Manual Control: Users can manually close notifications
- Debug Support: Console logging for troubleshooting
🎨 Enhanced User Interface
- Modern Chat Design: Sleek, intuitive chat interface with Scaffold AI branding
- Responsive Layout: Works seamlessly on desktop, tablet, and mobile devices
- Real-time Chat: Smooth message flow with typing indicators and proper error handling
- Reset Functionality: One-click conversation reset for easy testing and troubleshooting
- Debug Console: Comprehensive logging for developers and troubleshooting
📚 Intelligent Source Attribution
- Complete Source Information: Every response includes detailed source metadata:
- Document titles and authors
- Source folders and file paths
- Document IDs and chunk references
- DOIs when available
- Research-Based Responses: All recommendations backed by academic sustainability research
- Professional Formatting: Clean, structured responses with proper citations
🤖 LLM-Powered Intelligence
- Intelligent Response Generation: Contextual, nuanced responses using TinyLlama model
- Enhanced Prompt Engineering: Specialized prompts for sustainability education context
- Course-Specific Analysis: Intelligent analysis for different engineering disciplines:
- Fluid Mechanics integration strategies
- Thermodynamics sustainability approaches
- Materials science green technologies
- General sustainability principles
- Research Context Integration: LLM processes research data for informed recommendations
🔍 Advanced Search Capabilities
- Vector-Based Search: Semantic search through sustainability research database
- Hybrid Search System: Combines vector search with LLM processing
- Fallback Mechanisms: Robust error handling with graceful degradation
- Source Validation: Ensures all responses are backed by academic research
🚀 System Performance
- Hybrid Architecture: Combines LLM intelligence with vector search reliability
- Memory Optimization: Efficient resource usage and cleanup
- Production Ready: System tested and validated for extended use
- Scalable Design: Modular architecture for future enhancements
- Robust Error Handling: Graceful fallback when LLM is unavailable
🚀 Getting Started
Prerequisites
- Python 3.11 or higher (recommended: Python 3.11 for best compatibility)
- Git
- 16GB+ RAM recommended for optimal performance
- NVIDIA GPU recommended but not required
- Windows: Microsoft Visual C++ Build Tools (for some package installations)
- Linux: python3-dev and build-essential packages
Installation
Clone the repository:
bash git clone https://github.com/kevinmastascusa/scaffold_ai.git cd scaffold_aiCreate and activate virtual environment: ```bash
Windows (PowerShell)
Remove-Item -Path scaffoldenv -Recurse -Force -ErrorAction SilentlyContinue python -m venv scaffoldenv Set-ExecutionPolicy -ExecutionPolicy RemoteSigned -Scope Process .\scaffold_env\Scripts\activate
# macOS/Linux rm -rf scaffoldenv python -m venv scaffoldenv source scaffold_env/bin/activate ```
If you get a permission error, try:
- Running PowerShell as administrator (Windows)
- Using sudo for the commands (Linux/macOS)
- Checking if the directory is being used by another process
- Install dependencies:
bash # Upgrade pip first python -m pip install --upgrade pip # Install requirements (this may take 5-10 minutes) pip install -r requirements.txt
Note:
- Some packages are quite large (torch, transformers, etc.) and may take a while to download
- If you encounter issues:
- Windows: Install Visual C++ Build Tools
- Linux: Run sudo apt-get install python3-dev build-essential
- Run the setup script:
bash python setup.py
This will: - Create necessary directories if they don't exist (data/, outputs/, vectoroutputs/, mathoutputs/) - Validate the workspace structure - Check for existing PDF files - The script will work even if directories already exist
Configure Hugging Face (optional):
- For some models (like Llama 2), get your token from https://huggingface.co/settings/tokens
- Create a
.envfile:bash HUGGINGFACE_TOKEN=your_token_here
Add your PDF documents: ```bash
Create data directory if it doesn't exist (skip if it exists)
mkdir -p data ``
Place your PDF files in thedata/` directory. The system will automatically process all PDF files found in this directory and its subdirectories.
Troubleshooting
If you encounter issues during installation:
Virtual Environment Issues:
- Delete the existing
scaffold_envdirectory and try creating it again - Ensure you have write permissions in the current directory
- Try creating the virtual environment in a different location
- Delete the existing
Package Installation Errors:
- Make sure you're using Python 3.11 (some packages might not work with 3.12+)
- Install required system dependencies (Visual C++ Build Tools on Windows, build-essential on Linux)
- If a package fails to install, try installing it separately with
pip install package-name
Setup Script Errors:
- Ensure all directories are writable
- Check if any files are locked by other processes
- Make sure you're in the correct directory when running the script
Quick Start
- Process your documents: ```bash # Extract and chunk PDF documents python scaffold_core/scripts/chunk/ChunkTest.py
# Create vector embeddings python scaffold_core/vector/main.py ```
- Run analysis scripts: ```bash # Generate Unicode analysis report python scaffoldcore/scripts/tests/generateunicode_report.py
# Compare different extraction methods python scaffoldcore/scripts/tests/compareextractions.py
# Programmatically analyze and fix combined words
python scaffoldcore/scripts/utils/postprocesscombinedwords.py
python scaffoldcore/scripts/generatecombinedwordsreport.py
``
- Seeoutputs/combinedwordsanalysisreport.txt` for a detailed summary of remaining combined words (now mostly legitimate technical/academic terms).
- Test the query system (Added 6/29/2025): ```bash # Run comprehensive system tests python scaffoldcore/scripts/runtests.py
# Generate detailed test report
python scaffoldcore/scripts/generatetestreport.py
``
- Seedocumentation/querysystemtestreport.md` for comprehensive system analysis and test results.
Project Structure
scaffold_ai/
├── data/ # Place your PDF documents here
├── outputs/ # Chunked text extracts
├── vector_outputs/ # Vector embeddings and indexes
├── math_outputs/ # Math-aware processing results
├── scaffold_core/ # Core processing modules
│ ├── config.py # Central configuration
│ ├── llm.py # Hugging Face LLM integration
│ ├── scripts/ # Processing scripts
│ └── vector/ # Vector processing
├── setup.py # Setup script
└── requirements.txt # Python dependencies
Configuration
All paths and settings are centrally managed in scaffold_core/config.py. The configuration automatically adapts to your workspace location, making the project portable for anyone who clones the repository.
LLM Integration
The project uses Hugging Face's Transformers library with the following features: - Mistral-7B-Instruct model by default - Automatic GPU acceleration when available - Mixed precision for better memory efficiency - Easy model switching and parameter tuning
For detailed setup and configuration options, see the Local Setup Guide.
🎯 Goals and Objectives
The primary goal is to create a user-friendly, accurate, and literature-grounded AI tool capable of:
- 📚 Suggesting relevant and up-to-date curriculum content.
- 🔍 Ensuring transparency by referencing scholarly sources for every recommendation.
- 🧩 Facilitating easy integration into existing courses, supporting targeted learning outcomes.
🛠️ Combined Words Issue: Solved
- Combined words (e.g., "environmentalsustainability") are now programmatically detected and fixed.
- Automated post-processing and reporting scripts:
scaffold_core/scripts/utils/postprocess_combined_words.py(fixes combined words)scaffold_core/scripts/generate_combined_words_report.py(generates detailed report)
- Final analysis:
- No camelCase or PascalCase issues remain.
- Remaining combined words are legitimate technical/academic terms.
- See
outputs/combined_words_analysis_report.txtfor details.
🛠️ Proposed System Architecture
The system will include three key components:
- Retrieval-Augmented Generation (RAG) Framework
- Vector Embeddings: Pre-process and embed key sustainability and resilience literature into a vector database (e.g., FAISS, Pinecone).
- Document Retrieval: Efficiently search and retrieve relevant sections from scholarly sources based on embedded user queries.
🤖 Large Language Model (LLM)
Current Implementation (6/29/2025): * Mistral-7B-Instruct-v0.2 - Successfully integrated and tested - Official Mistral AI model with 7B parameters - Hugging Face Transformers integration with proper tokenization - CPU-based inference with mixed precision support - Comprehensive testing shows 100% success rate
Previously Considered Models: * Llama 3 (Meta): meta-llama/Llama-3.1-8B-Instruct, Llama-3.2-1B * Other Mistral Variants: Mistral-7B-v0.1, Mistral-7B-Instruct-v0.3 * Phi-3 Mini (Microsoft): Phi-3.5-mini-instruct, Phi-3-mini-4k-instruct
🔗 Citation Tracking and Transparency
- 🔗 Direct linking between generated content and original sources.
- 🖥️ Interactive UI to show how each recommendation is grounded in literature.
🔄 Technical Workflow
- 📥 Corpus Collection: Curate scholarly papers, reports, and policy documents.
- 🗃️ Data Preprocessing: Clean, segment, and prepare documents.
- 🤖 Embedding and Storage: Embed corpus data and store in a vector database.
- ⚙️ Inference Engine: Retrieve and use embeddings to augment LLM output.
- 📝 Citation Layer: Annotate outputs with clear citation links.
📅 Project Timeline Overview
The project follows a structured timeline with week-by-week development phases. Key phases include:
- 🏗️ Setting up the preprocessing pipeline and repository structure ✅
- 🤖 Embedding the curated document corpus and validating retrieval quality ✅
- 🧪 Integrating the LLM and developing the initial prototype ✅
- 🎨 Building and refining the user interface 🔄
- 🧾 Implementing citation tracking and performing usability testing 🔄
- 🧑🏫 Engaging stakeholders for feedback and refining the final product 📋
📊 For detailed project timeline and critical path analysis, see CRITICAL_PATH.md
Optional enhancements may include a real-time feedback loop in the UI and tag-based filtering of recommendations.
📈 Evaluation Overview
The system will be evaluated based on its ability to:
- 🤖 Retrieve relevant and accurate curriculum materials
- 🔍 Generate transparent, literature-backed recommendations
- ⚡ Provide a responsive and accessible user experience
- 👥 Satisfy stakeholders through iterative testing and feedback
Evaluation will include both qualitative feedback from faculty and technical performance benchmarks such as system responsiveness, citation traceability, and usability outcomes.
✅ Expected Outcomes
- 🛠️ A functioning prototype generating cited curriculum recommendations.
- 🖥️ Intuitive UI ready for pilot use.
- 📄 Comprehensive documentation for future development.
🧾 TODO Section
🔥 High Priority (Current Sprint)
Semantic Search and Retrieval ✅ COMPLETED (6/29/2025)
- ✅ Created query interface for the completed FAISS index
- ✅ Implemented semantic search functionality using vector embeddings
- ✅ Added comprehensive query performance testing and optimization
- ✅ Successfully integrated Mistral-7B-Instruct-v0.2 LLM
- ✅ All system components tested with 100% success rate
Citation Layer Implementation 🆕 NEW PRIORITY
- Implement automatic citation extraction and source linking
- Add citation formatting (APA, MLA, Chicago) and validation
- Display citations in LLM responses with proper attribution
Advanced Query Testing 🆕 NEW PRIORITY
- Create comprehensive sustainability query test suite
- Add query result quality assessment metrics
- Develop A/B testing framework for retrieval strategies
Build User Interface for Knowledge Base
- Develop web interface for querying the vectorized knowledge base
- Create intuitive search interface with result display
- Add query result visualization and export functionality
- Integrate citation display and source linking in UI
🔧 Medium Priority
Enhance PDF Extraction and Chunking
- Integrate full Unicode analysis into page-based chunks
- Complete math-aware chunking improvements in
ChunkTest_Math.py - Re-integrate math-aware chunking with vector pipeline
Advanced Search Features
- Implement filters and faceted search capabilities
- Add relevance scoring and ranking improvements
- Create advanced citation tracking and source linking
- Implement citation-based result ranking and credibility scoring
Testing and Validation ✅ LARGELY COMPLETED (6/29/2025)
- ✅ Implemented comprehensive automated testing system
- ✅ Added full test suite for vector operations and LLM integration
- ✅ Created performance benchmarks and validation metrics
- ✅ Generated detailed test reports with system specifications
- 🔄 Remaining: Configuration system validation tests
📈 Future Enhancements
Citation and Source Management
- Citation network analysis and bibliography generation
- Citation impact scoring and credibility metrics
- Citation export to reference management tools
Advanced Query Analytics
- Query intent classification and auto-completion
- Query result personalization and difficulty assessment
- Multi-modal query support (text + images + documents)
System Optimization
- Add support for incremental updates to the vector database
- Implement progress tracking for long-running processes
- Add configuration profiles for different deployment scenarios
Documentation and Examples
- Add vector query examples in the README
- Create user guides for different use cases
- Develop API documentation for programmatic access
Optional Enhancements
- Real-time feedback loop in the UI
- Tag-based filtering of recommendations
- Advanced analytics and usage reporting
📅 Week 1 Tasks
Define Preprocessing Methodology 🔄 LARGELY COMPLETE
- ✅ Established detailed document preprocessing methodology, including chunking size, format, and metadata extraction
- ✅ Implemented page-based chunking (one complete page per chunk)
- ✅ Created comprehensive Unicode and text analysis pipeline
- ✅ Processed 273 PDF documents successfully with 4,859 chunks generated
- 🔄 Still in progress: Math-aware chunking improvements and full Unicode integration
- 📝 Note: Methodology is largely defined but refinement may be needed based on downstream LLM integration and user feedback
GitHub Repository Setup ✅ COMPLETED
- ✅ Set up GitHub repository with appropriate structure, branches, and initial documentation
- ✅ Created centralized configuration system for portable deployment
- ✅ Implemented automated setup process with
python setup.py
Embedding Techniques and Vector Database ✅ COMPLETED
- ✅ Selected sentence-transformers for embedding generation
- ✅ Finalized FAISS as vector database choice
- ✅ Successfully generated embeddings for all 4,859 text chunks
- ✅ Created FAISS index for efficient similarity search and retrieval
- ✅ Resolved all dependency conflicts and compatibility issues
- ✅ Added 6/29/2025: Implemented full query system with LLM integration
Open-Source License Compliance 🔄 PENDING
- ✅ Confirmed that current libraries used (sentence-transformers, FAISS, torch) meet open-source license requirements
- 🔄 Pending: Final LLM model selection and license verification for downstream model integration
README.md Documentation ✅ COMPLETED
- ✅ Created and incrementally updated comprehensive README.md with full setup instructions, usage examples, and project context
- ✅ Added project structure overview and configuration explanation
🔍 Project Validation
To validate that all critical path components are present, run:
bash
python validate_critical_path.py
This will check all essential components and provide a status report.
📝 Model Version/Hash Logging
- Log all model names, descriptions, and hashes for reproducibility:
bash python -m scaffold_core.model_logging - See
outputs/model_version_log.jsonfor the log.
⚡ Model Benchmarking
- Benchmark all models for latency, memory, and output:
bash python -m scaffold_core.benchmark_models
Owner
- Login: kevinmastascusa
- Kind: user
- Repositories: 1
- Profile: https://github.com/kevinmastascusa
Citation (CITATION_DEBUG_REPORT.md)
# Citation and Relevance Debug Report
## 🎯 **Executive Summary**
The citation system is **functioning correctly** - all tests returned sources. However, there are **quality issues** with response relevance and repetition that need addressing.
## ✅ **What's Working**
1. **Vector Search**: ✅ Functioning correctly
- FAISS index: 4,859 vectors loaded
- Metadata: 4,859 entries available
- Candidates being found: 46-50 initial candidates per query
2. **Citation System**: ✅ **WORKING**
- All 3 test queries returned 3 sources each
- Source metadata includes: ID, name, file path, text preview
- Cross-encoder scoring is functioning
3. **UI Integration**: ✅ Working
- API endpoints responding correctly
- Search statistics being tracked
- Response generation working
## ⚠️ **Issues Identified**
### 1. **Response Quality Issues**
- **High Repetition Ratio**: 42-63% repetition in responses
- **Low Query Term Coverage**: 20% for first query (only 2/10 terms covered)
- **Response Length**: Varies significantly (261-736 words)
### 2. **Potential Citation Issues**
- **Score Variations**: Some sources have `nan` scores (Test 2)
- **Negative Scores**: Some sources have negative cross-encoder scores
- **Metadata Issues**: Some sources show "N/A" in direct vector search
## 🔍 **Root Cause Analysis**
### **Citation Issue (Teammate getting [])**
The system is **working correctly** for citations. If your teammate is getting empty arrays, possible causes:
1. **Different Query Processing**: Their queries might be hitting different code paths
2. **Filtering Logic**: There might be confidence thresholds filtering out valid citations
3. **Frontend Parsing**: The UI might not be displaying citations correctly
4. **Environment Differences**: Different Python versions or dependency versions
### **Relevance Issue**
The high repetition and low term coverage suggest:
1. **Prompt Engineering**: The system prompt may not be optimized for relevance
2. **Context Window**: Token limits may be truncating important context
3. **LLM Model**: TinyLlama might be generating repetitive responses
4. **Reranking**: Cross-encoder might not be effectively ranking for relevance
## 🛠️ **Recommended Fixes**
### **Immediate Actions**
1. **Check Teammate's Environment**:
```bash
# Verify Python version and dependencies
python --version
pip list | grep -E "(sentence-transformers|faiss|transformers)"
```
2. **Test Citation Filtering**:
```python
# Check if there are confidence thresholds
# Look for filtering logic in enhanced_query.py
```
3. **Verify Frontend Citation Display**:
```javascript
// Check if sources array is being parsed correctly in UI
```
### **Quality Improvements**
1. **Prompt Engineering**:
- Add explicit instructions for relevance and avoiding repetition
- Include query terms in the prompt more prominently
2. **Context Management**:
- Increase context window or improve chunking strategy
- Add query-specific context selection
3. **Reranking Optimization**:
- Adjust cross-encoder thresholds
- Implement better candidate filtering
## 📊 **Test Results Summary**
| Test | Query | Candidates | Sources | Repetition | Term Coverage |
|------|-------|------------|---------|------------|---------------|
| 1 | Sustainability in Fluid Mechanics | 3 | 3 | 46.4% | 20.0% |
| 2 | Climate Education Module | 3 | 3 | 63.5% | 83.3% |
| 3 | Critical Thinking Activity | 3 | 3 | 42.7% | 71.4% |
## 🎯 **Next Steps**
1. **For Citation Issue**:
- Have teammate run the debug script to compare results
- Check their environment and configuration
2. **For Quality Issue**:
- Review and optimize the system prompt
- Consider upgrading to a larger LLM model
- Implement better context selection
3. **Monitoring**:
- Add response quality metrics to production
- Implement citation confidence thresholds
## 📝 **Debug Commands**
```bash
# Run comprehensive debug
python debug_citation_issues.py
# Test specific query
curl -X POST -H "Content-Type: application/json" \
-d '{"query":"test query"}' \
http://localhost:5002/api/query
# Check vector search directly
python -c "
import sys; sys.path.append('/Users/kevinmastascusa/GITHUB/scaffold_ai')
from sentence_transformers import SentenceTransformer
import faiss
model = SentenceTransformer('all-MiniLM-L6-v2')
index = faiss.read_index('vector_outputs/scaffold_index_1.faiss')
query_embedding = model.encode(['test query'])
scores, indices = index.search(query_embedding, k=5)
print(f'Found {len(indices[0])} candidates')
"
```
---
**Status**: ✅ Citations working, ⚠️ Quality improvements needed
**Priority**: Medium (citations functional, focus on relevance)
**Assignee**: @Aethyrex
GitHub Events
Total
- Create event: 17
- Release event: 2
- Issues event: 12
- Watch event: 1
- Delete event: 7
- Issue comment event: 9
- Push event: 88
- Pull request review event: 1
- Gollum event: 2
- Pull request event: 24
Last Year
- Create event: 17
- Release event: 2
- Issues event: 12
- Watch event: 1
- Delete event: 7
- Issue comment event: 9
- Push event: 88
- Pull request review event: 1
- Gollum event: 2
- Pull request event: 24
Issues and Pull Requests
Last synced: 7 months ago
All Time
- Total issues: 0
- Total pull requests: 4
- Average time to close issues: N/A
- Average time to close pull requests: 6 days
- Total issue authors: 0
- Total pull request authors: 1
- Average comments per issue: 0
- Average comments per pull request: 1.0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 4
Past Year
- Issues: 0
- Pull requests: 4
- Average time to close issues: N/A
- Average time to close pull requests: 6 days
- Issue authors: 0
- Pull request authors: 1
- Average comments per issue: 0
- Average comments per pull request: 1.0
- Merged pull requests: 0
- Bot issues: 0
- Bot pull requests: 4
Top Authors
Issue Authors
- kevinmastascusa (8)
Pull Request Authors
- dependabot[bot] (10)
- Copilot (4)
- kevinmastascusa (1)
Top Labels
Issue Labels
Pull Request Labels
Dependencies
- faiss-cpu ==1.7.4
- fsspec ==2025.3.0
- huggingface-hub ==0.30.0
- mlflow ==2.5.0
- nltk ==3.8.1
- numpy ==1.26.0
- pyarrow <13,>=4.0.0
- sentence-transformers ==2.2.2
- torch *