GPTMB 2025 - The Second International Conference on Generative Pre-trained Transformer Models and Beyond
	July 06, 2025 - July 10, 2025
 GPTMB 2025
Onsite and Online Options: In order to accommodate various situations, we are offering the option for either physical presence or virtual participation (pdf slides or pre-recorded videos).
	ISSN: 
	ISBN: 978-1-68558-287-6 
		GPTMB 2025 is colocated with the following events as part of DigiTech 2025 Congress:
		  - DIGITAL 2025, Advances on Societal Digital Transformation
 	  - IoTAI 2025, The Second International Conference on IoT-AI
 	  - GPTMB 2025, The Second International Conference on Generative Pre-trained Transformer Models and Beyond
 	  - AIMEDIA 2025, The First International Conference on AI-based Media Innovation
 	
		
	GPTMB 2025 Steering Committee
	
      
        
           | 
            | 
          Petre Dini  
            IARIA 
            USA/EU 
              
  | 
           | 
            | 
          Isaac Caicedo-Castro 
            University of Córdoba 
            Colombia 
              
  | 
        
        
           | 
            | 
          Tzung-Pei Hong 
            National University of  Kaohsiung 
            Taiwan 
              
  | 
           | 
            | 
          Stephan Böhm 
            RheinMain        University of Applied Sciences - Wiesbaden 
            Germany 
              
  | 
        
        
           | 
            | 
          Zhixiong Chen 
            Mercy College 
            USA 
              
  | 
           | 
            | 
           Joni Salminen 
            University of Vaasa 
            Finland 
              
  | 
        
        
           | 
            | 
          Christelle Scharff 
            Pace University 
            USA 
              
  | 
           | 
            | 
          Gerald Penn 
            University of Toronto 
            Canada 
              
  | 
        
        
           | 
            | 
          Konstantinos (Constantine) Kotropoulos 
            Aristotle University of Thessalonik 
            Greece 
              
  | 
           | 
            | 
            | 
        
      
 
	
      
        
           | 
            | 
          Petre Dini  
            IARIA 
            USA/EU 
              
  | 
        
        
           | 
            | 
          Isaac Caicedo-Castro 
            University of Córdoba 
            Colombia 
              
  | 
        
        
           | 
            | 
          Tzung-Pei Hong 
            National University of  Kaohsiung 
            Taiwan 
              
  | 
        
        
           | 
            | 
          Stephan Böhm 
            RheinMain        University of Applied Sciences - Wiesbaden 
            Germany 
              
  | 
        
        
           | 
            | 
          Zhixiong Chen 
            Mercy College 
            USA 
              
  | 
        
        
           | 
            | 
           Joni Salminen 
            University of Vaasa 
            Finland 
              
  | 
        
        
           | 
            | 
          Christelle Scharff 
            Pace University 
            USA 
              
  | 
        
        
           | 
            | 
          Gerald Penn 
            University of Toronto 
            Canada 
              
  | 
        
        
           | 
            | 
          Konstantinos (Constantine) Kotropoulos 
            Aristotle University of Thessalonik 
            Greece 
              
  | 
        
      
 
	 
	GPTMB 2025 conference tracks:
	Generative-AI basics
	  Generative pre-trained transformer (GPT) models
	  Transformer-based models and LLMs (Large Language Models)
	  Combination of GPT models and Reinforcement learning models
	  Creativity and originality in GPT-based tools
	  Taxonomy of context-based LLM training
	  Deep learning and LLMs
	  Retrieval augmented generation (RAG) and fine-tunning LLMs
	  LLM and Reinforcement Learning from Human Feedback (RLHF)
	  LLMs (autoregressive, retrieval-augmented, autoencoding, reinforcement learning, etc.)
    Computational resources forLLM raining and for  LLM-based applications
	LLMs
	  Large Language Models (LLM) taxonomy
Model characteristics (architecture, size, training data and        duration)
Building, training, and fine tuning LLMs
Performance (accuracy, latency, scalability)
Capabilities (content generation, translation, interactive)
Domain (medical, legal, financial, education, etc.)
Ethics and safeness (bias, fairness, filter, explainability)
Legal (data privacy, data exfiltration, copyright, licensing)
Challenges (integrations, mismatching, overfitting, underfitting,        hallucinations, interpretability, bias mitigation, ethics)
	LLM-based tools and applications
	  Challenging requirements on basic actions and core principles
Methods for optimized selection of model size and complexity
Fine-tuning and personalization mechanisms
Human interactions and actions alignment
Multimodal input/output capabilities (text with visual, audio, and      other data types)
Adaptive learning or continuous learning (training optimization,      context-awareness)
Range of languages and dialects, including regional expansion 
Scalability, understandability, and explainability 
Tools for software development, planning, workflows, coding, etc.
Applications on robotics, autonomous systems, and moving targets
Cross-interdisciplinary applications (finance, healthcare,      technology, etc.)
Discovery and advanced scientific research applications 
Computational requirements and energy consumption
Efficient techniques (quantization, pruning, etc.)
Reliability and security of LLM-based applications
Co-creation, open source, and global accessibility
Ethical considerations (bias mitigation, fairness, responsibility)
	Small-language models and tiny-language models
Architecture and design principles specific to small language models
Tiny language models for smartphones, IoT devices, edge devices, and      embedded systems
Tools for small languages models (DistilBERT, TinyBERT, MiniLM,      etc.)
Knowledge distillation, quantization, low latency, resource      optimization
Energy efficiency for FPGAs and specialized ASICs for model      deployment
Tiny language models for real-time translation apps and mobile-based      chatbots
Tiny languages and federated        learning for privacy
Small language models for vision        for multimodal applications
Hardware considerations (energy, quantization, pruning, etc.)
Tiny language models and hardware accelerators (GPUs, TPUs, and      ML-custom ASICs)
	Critical Issues on Input Data
	  Datasets: accuracy, granularity, precision, false/true negative/positive
	  Visible vs invisible (private, personalized) data
	  Data extrapolation
	  Output biases and biased Datasets
	  Sensitivity and specificity of Datasets
	  Fake and incorrect information
	  Volatile data
	  Time sensitive data
	  Critical Issues on Processing
	  Process truthfulness
	  Understability, Interpretability, and Explainability
	  Detect biases and incorrectness
	  Incorporate the interactive feedback
	  Incorporate corrections
	  Retrieval augmented generation (RAG) for LLM input
RLHF for LLM fine-tuning output
	Output quality
	  Output biases and biased Datasets
	  Sensitivity and specificity of Datasets
	  Context-aware output
	  Fine/Coarse text summarization
	  Quality of Data pre-evaluation (obsolete, incomplete, fake, noisy, etc.)
	  Validation of output
	  Detect and expalin hallucinations
Detect biased and incorrect summarization before spreading it
	Education and academic liability issues
	  Curricula revision for enbedding AI-based tools and methodolgies
	  User awareness on output trust-ability
	  Copyright infringements rules
	  Plagiarism and self-plagiarism tools
	  Ownership infringement
	  Mechanisms for reference verification
Dealing with hidden self-references
	Regulations and limitations
	  Regulations (licensing, testing, compliance-threshold, decentralized/centralize innovations)
	  Mitigate societal risks of GPT models
	  Capturing emotion and sentience
	  Lack of personalized (individual) memory and memories (past facts)
	  Lack of instant personalized thinking (personalized summarization)
	  Risk of GPTM-based decisions
	  AI awareness
AI-induced deskilling 
	Case studies with analysis and testing AI applications
	  Lesson learned with existing tools (ChatGPT, Bard AI, ChatSonic, etc.)
	  Predictive analytics in healthcare
	  Medical Diagnostics
	  Medical Imaging 
	  Pharmacology
	  AI-based therapy
	  AI-based finance
	  AI-based planning
	  AI-based decision
	  AI-based systems control
	  AI-based education
	  AI-based cyber security
	
Deadlines:
	
              Submission  |         Apr 14, 2025  |       
              Notification  |         May 07, 2025  |       
              Registration  |         May 18, 2025  |       
              Camera ready  |         Jun 01, 2025  |       
    
Deadlines differ for special tracks. Please consult the conference home page for special tracks Call for Papers (if any).