From bbd1f48b789749e57c16f214dd81628823b4cf5a Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Tue, 31 Mar 2026 22:50:07 -0700 Subject: [PATCH 01/15] test: use gpt-oss-20b to do some minor refactoring --- src/main.rs | 33 +++++++++++++++++++++------------ src/ui/grid.rs | 9 +++++++-- 2 files changed, 28 insertions(+), 14 deletions(-) diff --git a/src/main.rs b/src/main.rs index 58526be..1e544d7 100644 --- a/src/main.rs +++ b/src/main.rs @@ -26,6 +26,11 @@ use ratatui::{ use model::Model; use ui::app::{App, AppMode}; + +fn render_paragraph(f: &mut Frame, area: Rect, text: &str, style: Style) { + f.render_widget(Paragraph::new(text).style(style), area); +} + use ui::category_panel::CategoryPanel; use ui::formula_panel::FormulaPanel; use ui::grid::GridWidget; @@ -332,14 +337,14 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { let right = " ?:help :q quit "; let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); let line = format!("{title}{pad}{right}"); - f.render_widget( - Paragraph::new(line).style( - Style::default() - .fg(Color::Black) - .bg(Color::Blue) - .add_modifier(Modifier::BOLD), - ), + render_paragraph( + f, area, + &line, + Style::default() + .fg(Color::Black) + .bg(Color::Blue) + .add_modifier(Modifier::BOLD), ); } @@ -450,7 +455,7 @@ fn draw_status(f: &mut Frame, area: Rect, app: &App) { _ => Style::default().fg(Color::Black).bg(Color::DarkGray), }; - f.render_widget(Paragraph::new(line).style(badge_style), area); + render_paragraph(f, area, &line, badge_style); } fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { @@ -460,9 +465,11 @@ fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { "" }; let line = format!(":{buf}▌"); - f.render_widget( - Paragraph::new(line).style(Style::default().fg(Color::White).bg(Color::Black)), + render_paragraph( + f, area, + &line, + Style::default().fg(Color::White).bg(Color::Black), ); } @@ -484,9 +491,11 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { .title(" Export CSV — path (Esc cancel) "); let inner = block.inner(popup_area); f.render_widget(block, popup_area); - f.render_widget( - Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), + render_paragraph( + f, inner, + &format!("{buf}▌"), + Style::default().fg(Color::Green), ); } diff --git a/src/ui/grid.rs b/src/ui/grid.rs index 6277f50..ca871b9 100644 --- a/src/ui/grid.rs +++ b/src/ui/grid.rs @@ -291,7 +291,7 @@ impl<'a> GridWidget<'a> { }; let value = self.model.evaluate(&key); - let cell_str = format_value(value.as_ref(), fmt_comma, fmt_decimals); + let cell_str = format_cell_value(value.as_ref(), &view.number_format); let is_selected = ri == sel_row && ci == sel_col; let is_search_match = !self.search_query.is_empty() && cell_str @@ -435,7 +435,7 @@ impl<'a> Widget for GridWidget<'a> { } } -fn format_value(v: Option<&CellValue>, comma: bool, decimals: u8) -> String { +pub fn format_value(v: Option<&CellValue>, comma: bool, decimals: u8) -> String { match v { Some(CellValue::Number(n)) => format_f64(*n, comma, decimals), Some(CellValue::Text(s)) => s.clone(), @@ -443,6 +443,11 @@ fn format_value(v: Option<&CellValue>, comma: bool, decimals: u8) -> String { } } +fn format_cell_value(v: Option<&CellValue>, fmt: &str) -> String { + let (comma, decimals) = parse_number_format(fmt); + format_value(v, comma, decimals) +} + pub fn parse_number_format(fmt: &str) -> (bool, u8) { let comma = fmt.contains(','); let decimals = fmt -- 2.51.0 From f1e6e61bcaa503e7f78d9f6fa2a15de004517028 Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Tue, 31 Mar 2026 22:50:10 -0700 Subject: [PATCH 02/15] Revert "test: use gpt-oss-20b to do some minor refactoring" This reverts commit bbd1f48b789749e57c16f214dd81628823b4cf5a. --- src/main.rs | 33 ++++++++++++--------------------- src/ui/grid.rs | 9 ++------- 2 files changed, 14 insertions(+), 28 deletions(-) diff --git a/src/main.rs b/src/main.rs index 1e544d7..58526be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -26,11 +26,6 @@ use ratatui::{ use model::Model; use ui::app::{App, AppMode}; - -fn render_paragraph(f: &mut Frame, area: Rect, text: &str, style: Style) { - f.render_widget(Paragraph::new(text).style(style), area); -} - use ui::category_panel::CategoryPanel; use ui::formula_panel::FormulaPanel; use ui::grid::GridWidget; @@ -337,14 +332,14 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { let right = " ?:help :q quit "; let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); let line = format!("{title}{pad}{right}"); - render_paragraph( - f, + f.render_widget( + Paragraph::new(line).style( + Style::default() + .fg(Color::Black) + .bg(Color::Blue) + .add_modifier(Modifier::BOLD), + ), area, - &line, - Style::default() - .fg(Color::Black) - .bg(Color::Blue) - .add_modifier(Modifier::BOLD), ); } @@ -455,7 +450,7 @@ fn draw_status(f: &mut Frame, area: Rect, app: &App) { _ => Style::default().fg(Color::Black).bg(Color::DarkGray), }; - render_paragraph(f, area, &line, badge_style); + f.render_widget(Paragraph::new(line).style(badge_style), area); } fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { @@ -465,11 +460,9 @@ fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { "" }; let line = format!(":{buf}▌"); - render_paragraph( - f, + f.render_widget( + Paragraph::new(line).style(Style::default().fg(Color::White).bg(Color::Black)), area, - &line, - Style::default().fg(Color::White).bg(Color::Black), ); } @@ -491,11 +484,9 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { .title(" Export CSV — path (Esc cancel) "); let inner = block.inner(popup_area); f.render_widget(block, popup_area); - render_paragraph( - f, + f.render_widget( + Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner, - &format!("{buf}▌"), - Style::default().fg(Color::Green), ); } diff --git a/src/ui/grid.rs b/src/ui/grid.rs index ca871b9..6277f50 100644 --- a/src/ui/grid.rs +++ b/src/ui/grid.rs @@ -291,7 +291,7 @@ impl<'a> GridWidget<'a> { }; let value = self.model.evaluate(&key); - let cell_str = format_cell_value(value.as_ref(), &view.number_format); + let cell_str = format_value(value.as_ref(), fmt_comma, fmt_decimals); let is_selected = ri == sel_row && ci == sel_col; let is_search_match = !self.search_query.is_empty() && cell_str @@ -435,7 +435,7 @@ impl<'a> Widget for GridWidget<'a> { } } -pub fn format_value(v: Option<&CellValue>, comma: bool, decimals: u8) -> String { +fn format_value(v: Option<&CellValue>, comma: bool, decimals: u8) -> String { match v { Some(CellValue::Number(n)) => format_f64(*n, comma, decimals), Some(CellValue::Text(s)) => s.clone(), @@ -443,11 +443,6 @@ pub fn format_value(v: Option<&CellValue>, comma: bool, decimals: u8) -> String } } -fn format_cell_value(v: Option<&CellValue>, fmt: &str) -> String { - let (comma, decimals) = parse_number_format(fmt); - format_value(v, comma, decimals) -} - pub fn parse_number_format(fmt: &str) -> (bool, u8) { let comma = fmt.contains(','); let decimals = fmt -- 2.51.0 From 98d151f345e96cbad77016055fb975c6be79df3b Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Tue, 31 Mar 2026 23:10:52 -0700 Subject: [PATCH 03/15] refactor: unsloth/Qwen3.5-35B-A3B-GGUF:Q5_K_M --- src/main.rs | 227 ++++++++++++++++++++++++++++------------------------ 1 file changed, 122 insertions(+), 105 deletions(-) diff --git a/src/main.rs b/src/main.rs index 58526be..c20de04 100644 --- a/src/main.rs +++ b/src/main.rs @@ -295,28 +295,10 @@ fn draw(f: &mut Frame, app: &App) { draw_title(f, main_chunks[0], app); draw_content(f, main_chunks[1], app); draw_tile_bar(f, main_chunks[2], app); - - if is_cmd_mode { - draw_command_bar(f, main_chunks[3], app); - } else { - draw_status(f, main_chunks[3], app); - } + draw_bottom_bar(f, main_chunks[3], app, is_cmd_mode); // Overlays (rendered last so they appear on top) - if matches!(app.mode, AppMode::Help) { - f.render_widget(HelpWidget, size); - } - if matches!(app.mode, AppMode::ImportWizard) { - if let Some(wizard) = &app.wizard { - f.render_widget(ImportWizardWidget::new(wizard), size); - } - } - if matches!(app.mode, AppMode::ExportPrompt { .. }) { - draw_export_prompt(f, size, app); - } - if app.is_empty_model() && matches!(app.mode, AppMode::Normal | AppMode::CommandMode { .. }) { - draw_welcome(f, main_chunks[1], app); - } + draw_overlays(f, main_chunks[1], app); } fn draw_title(f: &mut Frame, area: Rect, app: &App) { @@ -330,17 +312,8 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { .unwrap_or_default(); let title = format!(" improvise · {}{}{} ", app.model.name, file, dirty); let right = " ?:help :q quit "; - let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); - let line = format!("{title}{pad}{right}"); - f.render_widget( - Paragraph::new(line).style( - Style::default() - .fg(Color::Black) - .bg(Color::Blue) - .add_modifier(Modifier::BOLD), - ), - area, - ); + let line = fill_line(title, right, area.width); + f.render_widget(Paragraph::new(line).style(title_bar_style()), area); } fn draw_content(f: &mut Frame, area: Rect, app: &App) { @@ -406,51 +379,22 @@ fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { } fn draw_status(f: &mut Frame, area: Rect, app: &App) { - let mode_badge = match &app.mode { - AppMode::Normal => "NORMAL", - AppMode::Editing { .. } => "INSERT", - AppMode::FormulaEdit { .. } => "FORMULA", - AppMode::FormulaPanel => "FORMULAS", - AppMode::CategoryPanel => "CATEGORIES", - AppMode::CategoryAdd { .. } => "NEW CATEGORY", - AppMode::ItemAdd { .. } => "ADD ITEMS", - AppMode::ViewPanel => "VIEWS", - AppMode::TileSelect { .. } => "TILES", - AppMode::ImportWizard => "IMPORT", - AppMode::ExportPrompt { .. } => "EXPORT", - AppMode::CommandMode { .. } => "COMMAND", - AppMode::Help => "HELP", - AppMode::Quit => "QUIT", - }; - - let search_part = if app.search_mode { - format!(" /{}▌", app.search_query) - } else { - String::new() - }; - + let mode_badge = mode_badge_text(&app.mode); + let search_part = search_indicator(&app.search_query, app.search_mode); let msg = if !app.status_msg.is_empty() { app.status_msg.as_str() } else { app.hint_text() }; - let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); let left = format!(" {mode_badge}{search_part} {msg}"); - let right = view_badge; - let pad = " ".repeat((area.width as usize).saturating_sub(left.len() + right.len())); - let line = format!("{left}{pad}{right}"); - - let badge_style = match &app.mode { - AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), - AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), - AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), - _ => Style::default().fg(Color::Black).bg(Color::DarkGray), - }; - - f.render_widget(Paragraph::new(line).style(badge_style), area); + let line = fill_line(left, &view_badge, area.width); + f.render_widget( + Paragraph::new(line).style(status_bar_style(&app.mode)), + area, + ); } fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { @@ -460,10 +404,7 @@ fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { "" }; let line = format!(":{buf}▌"); - f.render_widget( - Paragraph::new(line).style(Style::default().fg(Color::White).bg(Color::Black)), - area, - ); + f.render_widget(Paragraph::new(line).style(command_bar_style()), area); } fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { @@ -472,18 +413,15 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { } else { "" }; - let popup_w = 64u16.min(area.width); - let x = area.x + area.width.saturating_sub(popup_w) / 2; - let y = area.y + area.height / 2; - let popup_area = Rect::new(x, y, popup_w, 3); + let popup = centered_popup(area, 64, 3); + f.render_widget(Clear, popup); - f.render_widget(Clear, popup_area); let block = Block::default() .borders(Borders::ALL) .border_style(Style::default().fg(Color::Yellow)) .title(" Export CSV — path (Esc cancel) "); - let inner = block.inner(popup_area); - f.render_widget(block, popup_area); + let inner = block.inner(popup); + f.render_widget(block, popup); f.render_widget( Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner, @@ -491,12 +429,7 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { } fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { - let w = 58u16.min(area.width.saturating_sub(4)); - let h = 20u16.min(area.height.saturating_sub(2)); - let x = area.x + area.width.saturating_sub(w) / 2; - let y = area.y + area.height.saturating_sub(h) / 2; - let popup = Rect::new(x, y, w, h); - + let popup = centered_popup(area, 58, 20); f.render_widget(Clear, popup); let block = Block::default() @@ -506,7 +439,35 @@ fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { let inner = block.inner(popup); f.render_widget(block, popup); - let lines: &[(&str, Style)] = &[ + let lines = welcome_lines(); + for (i, (text, style)) in lines.iter().enumerate() { + if i >= inner.height as usize { + break; + } + f.render_widget( + Paragraph::new(*text).style(*style), + Rect::new( + inner.x + 1, + inner.y + i as u16, + inner.width.saturating_sub(2), + 1, + ), + ); + } +} + +// ── Drawing Helpers ────────────────────────────────────────────────────────── + +fn title_bar_style() -> Style { + Style::default().fg(Color::Black).bg(Color::Blue).add_modifier(Modifier::BOLD) +} + +fn command_bar_style() -> Style { + Style::default().fg(Color::White).bg(Color::Black) +} + +fn welcome_lines() -> Vec<(&'static str, Style)> { + vec![ ( "Multi-dimensional data modeling — in your terminal.", Style::default().fg(Color::White), @@ -514,9 +475,7 @@ fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { ("", Style::default()), ( "Getting started", - Style::default() - .fg(Color::Blue) - .add_modifier(Modifier::BOLD), + Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD), ), ("", Style::default()), ( @@ -546,9 +505,7 @@ fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { ("", Style::default()), ( "Navigation", - Style::default() - .fg(Color::Blue) - .add_modifier(Modifier::BOLD), + Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD), ), ("", Style::default()), ( @@ -569,20 +526,80 @@ fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { Style::default(), ), (":q Quit", Style::default()), - ]; + ] +} - for (i, (text, style)) in lines.iter().enumerate() { - if i >= inner.height as usize { - break; - } - f.render_widget( - Paragraph::new(*text).style(*style), - Rect::new( - inner.x + 1, - inner.y + i as u16, - inner.width.saturating_sub(2), - 1, - ), - ); +fn fill_line(left: String, right: &str, width: u16) -> String { + let pad = " ".repeat((width as usize).saturating_sub(left.len() + right.len())); + format!("{left}{pad}{right}") +} + +fn status_bar_style(mode: &AppMode) -> Style { + match mode { + AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), + AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), + AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), + _ => Style::default().fg(Color::Black).bg(Color::DarkGray), + } +} + +fn mode_badge_text(mode: &AppMode) -> &'static str { + match mode { + AppMode::Normal => "NORMAL", + AppMode::Editing { .. } => "INSERT", + AppMode::FormulaEdit { .. } => "FORMULA", + AppMode::FormulaPanel => "FORMULAS", + AppMode::CategoryPanel => "CATEGORIES", + AppMode::CategoryAdd { .. } => "NEW CATEGORY", + AppMode::ItemAdd { .. } => "ADD ITEMS", + AppMode::ViewPanel => "VIEWS", + AppMode::TileSelect { .. } => "TILES", + AppMode::ImportWizard => "IMPORT", + AppMode::ExportPrompt { .. } => "EXPORT", + AppMode::CommandMode { .. } => "COMMAND", + AppMode::Help => "HELP", + AppMode::Quit => "QUIT", + } +} + +fn search_indicator(query: &str, active: bool) -> String { + if active { + format!(" /{}▌", query) + } else { + String::new() + } +} + +fn centered_popup(area: Rect, width: u16, height: u16) -> Rect { + let w = width.min(area.width); + let h = height.min(area.height); + let x = area.x + area.width.saturating_sub(w) / 2; + let y = area.y + area.height.saturating_sub(h) / 2; + Rect::new(x, y, w, h) +} + +fn draw_bottom_bar(f: &mut Frame, area: Rect, app: &App, is_cmd_mode: bool) { + if is_cmd_mode { + draw_command_bar(f, area, app); + } else { + draw_status(f, area, app); + } +} + +fn draw_overlays(f: &mut Frame, content_area: Rect, app: &App) { + let size = f.area(); + if matches!(app.mode, AppMode::Help) { + f.render_widget(HelpWidget, size); + } + if matches!(app.mode, AppMode::ImportWizard) { + if let Some(wizard) = &app.wizard { + f.render_widget(ImportWizardWidget::new(wizard), size); + } + } + if matches!(app.mode, AppMode::ExportPrompt { .. }) { + draw_export_prompt(f, size, app); + } + if app.is_empty_model() && matches!(app.mode, AppMode::Normal | AppMode::CommandMode { .. }) { + draw_welcome(f, content_area, app); } } -- 2.51.0 From 7731c7ceab8936bb6021386ee48eec9c09a5bee7 Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Tue, 31 Mar 2026 23:11:21 -0700 Subject: [PATCH 04/15] Revert "refactor: unsloth/Qwen3.5-35B-A3B-GGUF:Q5_K_M" This reverts commit 98d151f345e96cbad77016055fb975c6be79df3b. --- src/main.rs | 225 ++++++++++++++++++++++++---------------------------- 1 file changed, 104 insertions(+), 121 deletions(-) diff --git a/src/main.rs b/src/main.rs index c20de04..58526be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -295,10 +295,28 @@ fn draw(f: &mut Frame, app: &App) { draw_title(f, main_chunks[0], app); draw_content(f, main_chunks[1], app); draw_tile_bar(f, main_chunks[2], app); - draw_bottom_bar(f, main_chunks[3], app, is_cmd_mode); + + if is_cmd_mode { + draw_command_bar(f, main_chunks[3], app); + } else { + draw_status(f, main_chunks[3], app); + } // Overlays (rendered last so they appear on top) - draw_overlays(f, main_chunks[1], app); + if matches!(app.mode, AppMode::Help) { + f.render_widget(HelpWidget, size); + } + if matches!(app.mode, AppMode::ImportWizard) { + if let Some(wizard) = &app.wizard { + f.render_widget(ImportWizardWidget::new(wizard), size); + } + } + if matches!(app.mode, AppMode::ExportPrompt { .. }) { + draw_export_prompt(f, size, app); + } + if app.is_empty_model() && matches!(app.mode, AppMode::Normal | AppMode::CommandMode { .. }) { + draw_welcome(f, main_chunks[1], app); + } } fn draw_title(f: &mut Frame, area: Rect, app: &App) { @@ -312,8 +330,17 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { .unwrap_or_default(); let title = format!(" improvise · {}{}{} ", app.model.name, file, dirty); let right = " ?:help :q quit "; - let line = fill_line(title, right, area.width); - f.render_widget(Paragraph::new(line).style(title_bar_style()), area); + let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); + let line = format!("{title}{pad}{right}"); + f.render_widget( + Paragraph::new(line).style( + Style::default() + .fg(Color::Black) + .bg(Color::Blue) + .add_modifier(Modifier::BOLD), + ), + area, + ); } fn draw_content(f: &mut Frame, area: Rect, app: &App) { @@ -379,22 +406,51 @@ fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { } fn draw_status(f: &mut Frame, area: Rect, app: &App) { - let mode_badge = mode_badge_text(&app.mode); - let search_part = search_indicator(&app.search_query, app.search_mode); + let mode_badge = match &app.mode { + AppMode::Normal => "NORMAL", + AppMode::Editing { .. } => "INSERT", + AppMode::FormulaEdit { .. } => "FORMULA", + AppMode::FormulaPanel => "FORMULAS", + AppMode::CategoryPanel => "CATEGORIES", + AppMode::CategoryAdd { .. } => "NEW CATEGORY", + AppMode::ItemAdd { .. } => "ADD ITEMS", + AppMode::ViewPanel => "VIEWS", + AppMode::TileSelect { .. } => "TILES", + AppMode::ImportWizard => "IMPORT", + AppMode::ExportPrompt { .. } => "EXPORT", + AppMode::CommandMode { .. } => "COMMAND", + AppMode::Help => "HELP", + AppMode::Quit => "QUIT", + }; + + let search_part = if app.search_mode { + format!(" /{}▌", app.search_query) + } else { + String::new() + }; + let msg = if !app.status_msg.is_empty() { app.status_msg.as_str() } else { app.hint_text() }; + let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); let left = format!(" {mode_badge}{search_part} {msg}"); - let line = fill_line(left, &view_badge, area.width); - f.render_widget( - Paragraph::new(line).style(status_bar_style(&app.mode)), - area, - ); + let right = view_badge; + let pad = " ".repeat((area.width as usize).saturating_sub(left.len() + right.len())); + let line = format!("{left}{pad}{right}"); + + let badge_style = match &app.mode { + AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), + AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), + AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), + _ => Style::default().fg(Color::Black).bg(Color::DarkGray), + }; + + f.render_widget(Paragraph::new(line).style(badge_style), area); } fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { @@ -404,7 +460,10 @@ fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { "" }; let line = format!(":{buf}▌"); - f.render_widget(Paragraph::new(line).style(command_bar_style()), area); + f.render_widget( + Paragraph::new(line).style(Style::default().fg(Color::White).bg(Color::Black)), + area, + ); } fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { @@ -413,15 +472,18 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { } else { "" }; - let popup = centered_popup(area, 64, 3); - f.render_widget(Clear, popup); + let popup_w = 64u16.min(area.width); + let x = area.x + area.width.saturating_sub(popup_w) / 2; + let y = area.y + area.height / 2; + let popup_area = Rect::new(x, y, popup_w, 3); + f.render_widget(Clear, popup_area); let block = Block::default() .borders(Borders::ALL) .border_style(Style::default().fg(Color::Yellow)) .title(" Export CSV — path (Esc cancel) "); - let inner = block.inner(popup); - f.render_widget(block, popup); + let inner = block.inner(popup_area); + f.render_widget(block, popup_area); f.render_widget( Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner, @@ -429,7 +491,12 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { } fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { - let popup = centered_popup(area, 58, 20); + let w = 58u16.min(area.width.saturating_sub(4)); + let h = 20u16.min(area.height.saturating_sub(2)); + let x = area.x + area.width.saturating_sub(w) / 2; + let y = area.y + area.height.saturating_sub(h) / 2; + let popup = Rect::new(x, y, w, h); + f.render_widget(Clear, popup); let block = Block::default() @@ -439,35 +506,7 @@ fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { let inner = block.inner(popup); f.render_widget(block, popup); - let lines = welcome_lines(); - for (i, (text, style)) in lines.iter().enumerate() { - if i >= inner.height as usize { - break; - } - f.render_widget( - Paragraph::new(*text).style(*style), - Rect::new( - inner.x + 1, - inner.y + i as u16, - inner.width.saturating_sub(2), - 1, - ), - ); - } -} - -// ── Drawing Helpers ────────────────────────────────────────────────────────── - -fn title_bar_style() -> Style { - Style::default().fg(Color::Black).bg(Color::Blue).add_modifier(Modifier::BOLD) -} - -fn command_bar_style() -> Style { - Style::default().fg(Color::White).bg(Color::Black) -} - -fn welcome_lines() -> Vec<(&'static str, Style)> { - vec![ + let lines: &[(&str, Style)] = &[ ( "Multi-dimensional data modeling — in your terminal.", Style::default().fg(Color::White), @@ -475,7 +514,9 @@ fn welcome_lines() -> Vec<(&'static str, Style)> { ("", Style::default()), ( "Getting started", - Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD), + Style::default() + .fg(Color::Blue) + .add_modifier(Modifier::BOLD), ), ("", Style::default()), ( @@ -505,7 +546,9 @@ fn welcome_lines() -> Vec<(&'static str, Style)> { ("", Style::default()), ( "Navigation", - Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD), + Style::default() + .fg(Color::Blue) + .add_modifier(Modifier::BOLD), ), ("", Style::default()), ( @@ -526,80 +569,20 @@ fn welcome_lines() -> Vec<(&'static str, Style)> { Style::default(), ), (":q Quit", Style::default()), - ] -} + ]; -fn fill_line(left: String, right: &str, width: u16) -> String { - let pad = " ".repeat((width as usize).saturating_sub(left.len() + right.len())); - format!("{left}{pad}{right}") -} - -fn status_bar_style(mode: &AppMode) -> Style { - match mode { - AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), - AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), - AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), - _ => Style::default().fg(Color::Black).bg(Color::DarkGray), - } -} - -fn mode_badge_text(mode: &AppMode) -> &'static str { - match mode { - AppMode::Normal => "NORMAL", - AppMode::Editing { .. } => "INSERT", - AppMode::FormulaEdit { .. } => "FORMULA", - AppMode::FormulaPanel => "FORMULAS", - AppMode::CategoryPanel => "CATEGORIES", - AppMode::CategoryAdd { .. } => "NEW CATEGORY", - AppMode::ItemAdd { .. } => "ADD ITEMS", - AppMode::ViewPanel => "VIEWS", - AppMode::TileSelect { .. } => "TILES", - AppMode::ImportWizard => "IMPORT", - AppMode::ExportPrompt { .. } => "EXPORT", - AppMode::CommandMode { .. } => "COMMAND", - AppMode::Help => "HELP", - AppMode::Quit => "QUIT", - } -} - -fn search_indicator(query: &str, active: bool) -> String { - if active { - format!(" /{}▌", query) - } else { - String::new() - } -} - -fn centered_popup(area: Rect, width: u16, height: u16) -> Rect { - let w = width.min(area.width); - let h = height.min(area.height); - let x = area.x + area.width.saturating_sub(w) / 2; - let y = area.y + area.height.saturating_sub(h) / 2; - Rect::new(x, y, w, h) -} - -fn draw_bottom_bar(f: &mut Frame, area: Rect, app: &App, is_cmd_mode: bool) { - if is_cmd_mode { - draw_command_bar(f, area, app); - } else { - draw_status(f, area, app); - } -} - -fn draw_overlays(f: &mut Frame, content_area: Rect, app: &App) { - let size = f.area(); - if matches!(app.mode, AppMode::Help) { - f.render_widget(HelpWidget, size); - } - if matches!(app.mode, AppMode::ImportWizard) { - if let Some(wizard) = &app.wizard { - f.render_widget(ImportWizardWidget::new(wizard), size); + for (i, (text, style)) in lines.iter().enumerate() { + if i >= inner.height as usize { + break; } - } - if matches!(app.mode, AppMode::ExportPrompt { .. }) { - draw_export_prompt(f, size, app); - } - if app.is_empty_model() && matches!(app.mode, AppMode::Normal | AppMode::CommandMode { .. }) { - draw_welcome(f, content_area, app); + f.render_widget( + Paragraph::new(*text).style(*style), + Rect::new( + inner.x + 1, + inner.y + i as u16, + inner.width.saturating_sub(2), + 1, + ), + ); } } -- 2.51.0 From d91590835424cc104db23bc45c965aff103a32ce Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:20:19 -0700 Subject: [PATCH 05/15] refactor: unsloth/Qwen3-Coder-Next-GGUF:Q5_K_M refactors the drawing helper --- src/main.rs | 328 +++++++++++++++++----------------------------------- 1 file changed, 104 insertions(+), 224 deletions(-) diff --git a/src/main.rs b/src/main.rs index 58526be..607d42f 100644 --- a/src/main.rs +++ b/src/main.rs @@ -280,8 +280,6 @@ fn run_tui( fn draw(f: &mut Frame, app: &App) { let size = f.area(); - let is_cmd_mode = matches!(app.mode, AppMode::CommandMode { .. }); - let main_chunks = Layout::default() .direction(Direction::Vertical) .constraints([ @@ -295,12 +293,7 @@ fn draw(f: &mut Frame, app: &App) { draw_title(f, main_chunks[0], app); draw_content(f, main_chunks[1], app); draw_tile_bar(f, main_chunks[2], app); - - if is_cmd_mode { - draw_command_bar(f, main_chunks[3], app); - } else { - draw_status(f, main_chunks[3], app); - } + draw_bottom_bar(f, main_chunks[3], app); // Overlays (rendered last so they appear on top) if matches!(app.mode, AppMode::Help) { @@ -312,92 +305,48 @@ fn draw(f: &mut Frame, app: &App) { } } if matches!(app.mode, AppMode::ExportPrompt { .. }) { - draw_export_prompt(f, size, app); + draw_export_prompt(f, app); } if app.is_empty_model() && matches!(app.mode, AppMode::Normal | AppMode::CommandMode { .. }) { - draw_welcome(f, main_chunks[1], app); + draw_welcome(f, main_chunks[1]); } } fn draw_title(f: &mut Frame, area: Rect, app: &App) { let dirty = if app.dirty { " [+]" } else { "" }; - let file = app - .file_path - .as_ref() - .and_then(|p| p.file_name()) - .and_then(|n| n.to_str()) - .map(|n| format!(" ({n})")) - .unwrap_or_default(); + let file = app.file_path.as_ref().and_then(|p| p.file_name()).and_then(|n| n.to_str()).map(|n| format!(" ({n})")).unwrap_or_default(); let title = format!(" improvise · {}{}{} ", app.model.name, file, dirty); let right = " ?:help :q quit "; let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); let line = format!("{title}{pad}{right}"); - f.render_widget( - Paragraph::new(line).style( - Style::default() - .fg(Color::Black) - .bg(Color::Blue) - .add_modifier(Modifier::BOLD), - ), - area, - ); + f.render_widget(Paragraph::new(line).style(Style::default().fg(Color::Black).bg(Color::Blue).add_modifier(Modifier::BOLD)), area); } fn draw_content(f: &mut Frame, area: Rect, app: &App) { - let side_open = app.formula_panel_open || app.category_panel_open || app.view_panel_open; - - if side_open { + if app.formula_panel_open || app.category_panel_open || app.view_panel_open { let side_w = 32u16; - let chunks = Layout::default() - .direction(Direction::Horizontal) - .constraints([Constraint::Min(40), Constraint::Length(side_w)]) - .split(area); - - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - chunks[0], - ); + let chunks = Layout::default().direction(Direction::Horizontal).constraints([Constraint::Min(40), Constraint::Length(side_w)]).split(area); + f.render_widget(GridWidget::new(&app.model, &app.mode, &app.search_query), chunks[0]); let side = chunks[1]; - let panel_count = [ - app.formula_panel_open, - app.category_panel_open, - app.view_panel_open, - ] - .iter() - .filter(|&&b| b) - .count() as u16; + let open_panels = [app.formula_panel_open, app.category_panel_open, app.view_panel_open]; + let panel_count = open_panels.iter().filter(|&&b| b).count() as u16; let ph = side.height / panel_count.max(1); let mut y = side.y; if app.formula_panel_open { - let a = Rect::new(side.x, y, side.width, ph); - f.render_widget( - FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), - a, - ); + f.render_widget(FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), Rect::new(side.x, y, side.width, ph)); y += ph; } if app.category_panel_open { - let a = Rect::new(side.x, y, side.width, ph); - f.render_widget( - CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), - a, - ); + f.render_widget(CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), Rect::new(side.x, y, side.width, ph)); y += ph; } if app.view_panel_open { - let a = Rect::new(side.x, y, side.width, ph); - f.render_widget( - ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), - a, - ); + f.render_widget(ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), Rect::new(side.x, y, side.width, ph)); } } else { - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - area, - ); + f.render_widget(GridWidget::new(&app.model, &app.mode, &app.search_query), area); } } @@ -405,8 +354,91 @@ fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { f.render_widget(TileBar::new(&app.model, &app.mode), area); } +fn draw_bottom_bar(f: &mut Frame, area: Rect, app: &App) { + match app.mode { + AppMode::CommandMode { ref buffer } => draw_command_bar(f, area, buffer), + _ => draw_status(f, area, app), + } +} + fn draw_status(f: &mut Frame, area: Rect, app: &App) { - let mode_badge = match &app.mode { + let mode_badge = mode_name(&app.mode); + let search_part = if app.search_mode { format!(" /{}▌", app.search_query) } else { String::new() }; + let msg = if !app.status_msg.is_empty() { app.status_msg.as_str() } else { app.hint_text() }; + let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; + let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); + let left = format!(" {mode_badge}{search_part} {msg}"); + let right = view_badge; + let pad = " ".repeat((area.width as usize).saturating_sub(left.len() + right.len())); + let line = format!("{left}{pad}{right}"); + let badge_style = mode_style(&app.mode); + f.render_widget(Paragraph::new(line).style(badge_style), area); +} + +fn draw_command_bar(f: &mut Frame, area: Rect, buffer: &str) { + f.render_widget(Paragraph::new(format!(":{buffer}▌")).style(Style::default().fg(Color::White).bg(Color::Black)), area); +} + +fn draw_export_prompt(f: &mut Frame, app: &App) { + let area = f.area(); + let popup_w = 64u16.min(area.width); + let x = area.x + area.width.saturating_sub(popup_w) / 2; + let y = area.y + area.height / 2; + let popup_area = Rect::new(x, y, popup_w, 3); + + f.render_widget(Clear, popup_area); + let block = Block::default().borders(Borders::ALL).border_style(Style::default().fg(Color::Yellow)).title(" Export CSV — path (Esc cancel) "); + let inner = block.inner(popup_area); + f.render_widget(block, popup_area); + let buf = if let AppMode::ExportPrompt { buffer } = &app.mode { buffer } else { "" }; + f.render_widget(Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner); +} + +fn draw_welcome(f: &mut Frame, area: Rect) { + let w = 58u16.min(area.width.saturating_sub(4)); + let h = 20u16.min(area.height.saturating_sub(2)); + let x = area.x + area.width.saturating_sub(w) / 2; + let y = area.y + area.height.saturating_sub(h) / 2; + let popup = Rect::new(x, y, w, h); + + f.render_widget(Clear, popup); + + let block = Block::default().borders(Borders::ALL).border_style(Style::default().fg(Color::Blue)).title(" Welcome to improvise "); + let inner = block.inner(popup); + f.render_widget(block, popup); + + let lines: &[(&str, Style)] = &[ + ("Multi-dimensional data modeling — in your terminal.", Style::default().fg(Color::White)), + ("", Style::default()), + ("Getting started", Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD)), + ("", Style::default()), + (":import Import a JSON file", Style::default().fg(Color::Cyan)), + (":add-cat Add a category (dimension)", Style::default().fg(Color::Cyan)), + (":add-item Add an item to a category", Style::default().fg(Color::Cyan)), + (":formula Add a formula, e.g.:", Style::default().fg(Color::Cyan)), + (" Profit = Revenue - Cost", Style::default().fg(Color::Green)), + (":w Save your model", Style::default().fg(Color::Cyan)), + ("", Style::default()), + ("Navigation", Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD)), + ("", Style::default()), + ("F C V Open panels (Formulas/Categories/Views)", Style::default()), + ("T Tile-select: pivot rows ↔ cols ↔ page", Style::default()), + ("i Enter Edit a cell", Style::default()), + ("[ ] Cycle the page-axis filter", Style::default()), + ("? or :help Full key reference", Style::default()), + (":q Quit", Style::default()), + ]; + + for (i, (text, style)) in lines.iter().enumerate() { + if i >= inner.height as usize { break; } + f.render_widget(Paragraph::new(*text).style(*style), Rect::new(inner.x + 1, inner.y + i as u16, inner.width.saturating_sub(2), 1)); + } +} + +// ── Helpers ────────────────────────────────────────────────────────────────── + +fn mode_name(mode: &AppMode) -> &'static str { + match mode { AppMode::Normal => "NORMAL", AppMode::Editing { .. } => "INSERT", AppMode::FormulaEdit { .. } => "FORMULA", @@ -421,168 +453,16 @@ fn draw_status(f: &mut Frame, area: Rect, app: &App) { AppMode::CommandMode { .. } => "COMMAND", AppMode::Help => "HELP", AppMode::Quit => "QUIT", - }; + } +} - let search_part = if app.search_mode { - format!(" /{}▌", app.search_query) - } else { - String::new() - }; - - let msg = if !app.status_msg.is_empty() { - app.status_msg.as_str() - } else { - app.hint_text() - }; - - let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; - let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); - - let left = format!(" {mode_badge}{search_part} {msg}"); - let right = view_badge; - let pad = " ".repeat((area.width as usize).saturating_sub(left.len() + right.len())); - let line = format!("{left}{pad}{right}"); - - let badge_style = match &app.mode { +fn mode_style(mode: &AppMode) -> Style { + match mode { AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), _ => Style::default().fg(Color::Black).bg(Color::DarkGray), - }; - - f.render_widget(Paragraph::new(line).style(badge_style), area); -} - -fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { - let buf = if let AppMode::CommandMode { buffer } = &app.mode { - buffer.as_str() - } else { - "" - }; - let line = format!(":{buf}▌"); - f.render_widget( - Paragraph::new(line).style(Style::default().fg(Color::White).bg(Color::Black)), - area, - ); -} - -fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { - let buf = if let AppMode::ExportPrompt { buffer } = &app.mode { - buffer.as_str() - } else { - "" - }; - let popup_w = 64u16.min(area.width); - let x = area.x + area.width.saturating_sub(popup_w) / 2; - let y = area.y + area.height / 2; - let popup_area = Rect::new(x, y, popup_w, 3); - - f.render_widget(Clear, popup_area); - let block = Block::default() - .borders(Borders::ALL) - .border_style(Style::default().fg(Color::Yellow)) - .title(" Export CSV — path (Esc cancel) "); - let inner = block.inner(popup_area); - f.render_widget(block, popup_area); - f.render_widget( - Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), - inner, - ); -} - -fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { - let w = 58u16.min(area.width.saturating_sub(4)); - let h = 20u16.min(area.height.saturating_sub(2)); - let x = area.x + area.width.saturating_sub(w) / 2; - let y = area.y + area.height.saturating_sub(h) / 2; - let popup = Rect::new(x, y, w, h); - - f.render_widget(Clear, popup); - - let block = Block::default() - .borders(Borders::ALL) - .border_style(Style::default().fg(Color::Blue)) - .title(" Welcome to improvise "); - let inner = block.inner(popup); - f.render_widget(block, popup); - - let lines: &[(&str, Style)] = &[ - ( - "Multi-dimensional data modeling — in your terminal.", - Style::default().fg(Color::White), - ), - ("", Style::default()), - ( - "Getting started", - Style::default() - .fg(Color::Blue) - .add_modifier(Modifier::BOLD), - ), - ("", Style::default()), - ( - ":import Import a JSON file", - Style::default().fg(Color::Cyan), - ), - ( - ":add-cat Add a category (dimension)", - Style::default().fg(Color::Cyan), - ), - ( - ":add-item Add an item to a category", - Style::default().fg(Color::Cyan), - ), - ( - ":formula Add a formula, e.g.:", - Style::default().fg(Color::Cyan), - ), - ( - " Profit = Revenue - Cost", - Style::default().fg(Color::Green), - ), - ( - ":w Save your model", - Style::default().fg(Color::Cyan), - ), - ("", Style::default()), - ( - "Navigation", - Style::default() - .fg(Color::Blue) - .add_modifier(Modifier::BOLD), - ), - ("", Style::default()), - ( - "F C V Open panels (Formulas/Categories/Views)", - Style::default(), - ), - ( - "T Tile-select: pivot rows ↔ cols ↔ page", - Style::default(), - ), - ("i Enter Edit a cell", Style::default()), - ( - "[ ] Cycle the page-axis filter", - Style::default(), - ), - ( - "? or :help Full key reference", - Style::default(), - ), - (":q Quit", Style::default()), - ]; - - for (i, (text, style)) in lines.iter().enumerate() { - if i >= inner.height as usize { - break; - } - f.render_widget( - Paragraph::new(*text).style(*style), - Rect::new( - inner.x + 1, - inner.y + i as u16, - inner.width.saturating_sub(2), - 1, - ), - ); } } + + -- 2.51.0 From 8c84256ebc4e5c7e1647e28d99fb01b007feca9c Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:25:19 -0700 Subject: [PATCH 06/15] refactor: merge using claude sonnet --- src/main.rs | 30 +++++++++++++++--------------- 1 file changed, 15 insertions(+), 15 deletions(-) diff --git a/src/main.rs b/src/main.rs index 607d42f..00b8c22 100644 --- a/src/main.rs +++ b/src/main.rs @@ -317,8 +317,7 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { let file = app.file_path.as_ref().and_then(|p| p.file_name()).and_then(|n| n.to_str()).map(|n| format!(" ({n})")).unwrap_or_default(); let title = format!(" improvise · {}{}{} ", app.model.name, file, dirty); let right = " ?:help :q quit "; - let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); - let line = format!("{title}{pad}{right}"); + let line = fill_line(title, right, area.width); f.render_widget(Paragraph::new(line).style(Style::default().fg(Color::Black).bg(Color::Blue).add_modifier(Modifier::BOLD)), area); } @@ -368,9 +367,7 @@ fn draw_status(f: &mut Frame, area: Rect, app: &App) { let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); let left = format!(" {mode_badge}{search_part} {msg}"); - let right = view_badge; - let pad = " ".repeat((area.width as usize).saturating_sub(left.len() + right.len())); - let line = format!("{left}{pad}{right}"); + let line = fill_line(left, &view_badge, area.width); let badge_style = mode_style(&app.mode); f.render_widget(Paragraph::new(line).style(badge_style), area); } @@ -380,11 +377,7 @@ fn draw_command_bar(f: &mut Frame, area: Rect, buffer: &str) { } fn draw_export_prompt(f: &mut Frame, app: &App) { - let area = f.area(); - let popup_w = 64u16.min(area.width); - let x = area.x + area.width.saturating_sub(popup_w) / 2; - let y = area.y + area.height / 2; - let popup_area = Rect::new(x, y, popup_w, 3); + let popup_area = centered_popup(f.area(), 64, 3); f.render_widget(Clear, popup_area); let block = Block::default().borders(Borders::ALL).border_style(Style::default().fg(Color::Yellow)).title(" Export CSV — path (Esc cancel) "); @@ -395,11 +388,7 @@ fn draw_export_prompt(f: &mut Frame, app: &App) { } fn draw_welcome(f: &mut Frame, area: Rect) { - let w = 58u16.min(area.width.saturating_sub(4)); - let h = 20u16.min(area.height.saturating_sub(2)); - let x = area.x + area.width.saturating_sub(w) / 2; - let y = area.y + area.height.saturating_sub(h) / 2; - let popup = Rect::new(x, y, w, h); + let popup = centered_popup(area, 58, 20); f.render_widget(Clear, popup); @@ -465,4 +454,15 @@ fn mode_style(mode: &AppMode) -> Style { } } +fn fill_line(left: String, right: &str, width: u16) -> String { + let pad = " ".repeat((width as usize).saturating_sub(left.len() + right.len())); + format!("{left}{pad}{right}") +} +fn centered_popup(area: Rect, width: u16, height: u16) -> Rect { + let w = width.min(area.width); + let h = height.min(area.height); + let x = area.x + area.width.saturating_sub(w) / 2; + let y = area.y + area.height.saturating_sub(h) / 2; + Rect::new(x, y, w, h) +} -- 2.51.0 From ff08e3c2c249021f763c1d3b62574fff55099224 Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:26:55 -0700 Subject: [PATCH 07/15] chore: Revert refactors to give claude a clean slate --- src/main.rs | 322 ++++++++++++++++++++++++++++++++++++---------------- 1 file changed, 221 insertions(+), 101 deletions(-) diff --git a/src/main.rs b/src/main.rs index 00b8c22..58526be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -280,6 +280,8 @@ fn run_tui( fn draw(f: &mut Frame, app: &App) { let size = f.area(); + let is_cmd_mode = matches!(app.mode, AppMode::CommandMode { .. }); + let main_chunks = Layout::default() .direction(Direction::Vertical) .constraints([ @@ -293,7 +295,12 @@ fn draw(f: &mut Frame, app: &App) { draw_title(f, main_chunks[0], app); draw_content(f, main_chunks[1], app); draw_tile_bar(f, main_chunks[2], app); - draw_bottom_bar(f, main_chunks[3], app); + + if is_cmd_mode { + draw_command_bar(f, main_chunks[3], app); + } else { + draw_status(f, main_chunks[3], app); + } // Overlays (rendered last so they appear on top) if matches!(app.mode, AppMode::Help) { @@ -305,47 +312,92 @@ fn draw(f: &mut Frame, app: &App) { } } if matches!(app.mode, AppMode::ExportPrompt { .. }) { - draw_export_prompt(f, app); + draw_export_prompt(f, size, app); } if app.is_empty_model() && matches!(app.mode, AppMode::Normal | AppMode::CommandMode { .. }) { - draw_welcome(f, main_chunks[1]); + draw_welcome(f, main_chunks[1], app); } } fn draw_title(f: &mut Frame, area: Rect, app: &App) { let dirty = if app.dirty { " [+]" } else { "" }; - let file = app.file_path.as_ref().and_then(|p| p.file_name()).and_then(|n| n.to_str()).map(|n| format!(" ({n})")).unwrap_or_default(); + let file = app + .file_path + .as_ref() + .and_then(|p| p.file_name()) + .and_then(|n| n.to_str()) + .map(|n| format!(" ({n})")) + .unwrap_or_default(); let title = format!(" improvise · {}{}{} ", app.model.name, file, dirty); let right = " ?:help :q quit "; - let line = fill_line(title, right, area.width); - f.render_widget(Paragraph::new(line).style(Style::default().fg(Color::Black).bg(Color::Blue).add_modifier(Modifier::BOLD)), area); + let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); + let line = format!("{title}{pad}{right}"); + f.render_widget( + Paragraph::new(line).style( + Style::default() + .fg(Color::Black) + .bg(Color::Blue) + .add_modifier(Modifier::BOLD), + ), + area, + ); } fn draw_content(f: &mut Frame, area: Rect, app: &App) { - if app.formula_panel_open || app.category_panel_open || app.view_panel_open { + let side_open = app.formula_panel_open || app.category_panel_open || app.view_panel_open; + + if side_open { let side_w = 32u16; - let chunks = Layout::default().direction(Direction::Horizontal).constraints([Constraint::Min(40), Constraint::Length(side_w)]).split(area); - f.render_widget(GridWidget::new(&app.model, &app.mode, &app.search_query), chunks[0]); + let chunks = Layout::default() + .direction(Direction::Horizontal) + .constraints([Constraint::Min(40), Constraint::Length(side_w)]) + .split(area); + + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + chunks[0], + ); let side = chunks[1]; - let open_panels = [app.formula_panel_open, app.category_panel_open, app.view_panel_open]; - let panel_count = open_panels.iter().filter(|&&b| b).count() as u16; + let panel_count = [ + app.formula_panel_open, + app.category_panel_open, + app.view_panel_open, + ] + .iter() + .filter(|&&b| b) + .count() as u16; let ph = side.height / panel_count.max(1); let mut y = side.y; if app.formula_panel_open { - f.render_widget(FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), Rect::new(side.x, y, side.width, ph)); + let a = Rect::new(side.x, y, side.width, ph); + f.render_widget( + FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), + a, + ); y += ph; } if app.category_panel_open { - f.render_widget(CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), Rect::new(side.x, y, side.width, ph)); + let a = Rect::new(side.x, y, side.width, ph); + f.render_widget( + CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), + a, + ); y += ph; } if app.view_panel_open { - f.render_widget(ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), Rect::new(side.x, y, side.width, ph)); + let a = Rect::new(side.x, y, side.width, ph); + f.render_widget( + ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), + a, + ); } } else { - f.render_widget(GridWidget::new(&app.model, &app.mode, &app.search_query), area); + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + area, + ); } } @@ -353,81 +405,8 @@ fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { f.render_widget(TileBar::new(&app.model, &app.mode), area); } -fn draw_bottom_bar(f: &mut Frame, area: Rect, app: &App) { - match app.mode { - AppMode::CommandMode { ref buffer } => draw_command_bar(f, area, buffer), - _ => draw_status(f, area, app), - } -} - fn draw_status(f: &mut Frame, area: Rect, app: &App) { - let mode_badge = mode_name(&app.mode); - let search_part = if app.search_mode { format!(" /{}▌", app.search_query) } else { String::new() }; - let msg = if !app.status_msg.is_empty() { app.status_msg.as_str() } else { app.hint_text() }; - let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; - let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); - let left = format!(" {mode_badge}{search_part} {msg}"); - let line = fill_line(left, &view_badge, area.width); - let badge_style = mode_style(&app.mode); - f.render_widget(Paragraph::new(line).style(badge_style), area); -} - -fn draw_command_bar(f: &mut Frame, area: Rect, buffer: &str) { - f.render_widget(Paragraph::new(format!(":{buffer}▌")).style(Style::default().fg(Color::White).bg(Color::Black)), area); -} - -fn draw_export_prompt(f: &mut Frame, app: &App) { - let popup_area = centered_popup(f.area(), 64, 3); - - f.render_widget(Clear, popup_area); - let block = Block::default().borders(Borders::ALL).border_style(Style::default().fg(Color::Yellow)).title(" Export CSV — path (Esc cancel) "); - let inner = block.inner(popup_area); - f.render_widget(block, popup_area); - let buf = if let AppMode::ExportPrompt { buffer } = &app.mode { buffer } else { "" }; - f.render_widget(Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner); -} - -fn draw_welcome(f: &mut Frame, area: Rect) { - let popup = centered_popup(area, 58, 20); - - f.render_widget(Clear, popup); - - let block = Block::default().borders(Borders::ALL).border_style(Style::default().fg(Color::Blue)).title(" Welcome to improvise "); - let inner = block.inner(popup); - f.render_widget(block, popup); - - let lines: &[(&str, Style)] = &[ - ("Multi-dimensional data modeling — in your terminal.", Style::default().fg(Color::White)), - ("", Style::default()), - ("Getting started", Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD)), - ("", Style::default()), - (":import Import a JSON file", Style::default().fg(Color::Cyan)), - (":add-cat Add a category (dimension)", Style::default().fg(Color::Cyan)), - (":add-item Add an item to a category", Style::default().fg(Color::Cyan)), - (":formula Add a formula, e.g.:", Style::default().fg(Color::Cyan)), - (" Profit = Revenue - Cost", Style::default().fg(Color::Green)), - (":w Save your model", Style::default().fg(Color::Cyan)), - ("", Style::default()), - ("Navigation", Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD)), - ("", Style::default()), - ("F C V Open panels (Formulas/Categories/Views)", Style::default()), - ("T Tile-select: pivot rows ↔ cols ↔ page", Style::default()), - ("i Enter Edit a cell", Style::default()), - ("[ ] Cycle the page-axis filter", Style::default()), - ("? or :help Full key reference", Style::default()), - (":q Quit", Style::default()), - ]; - - for (i, (text, style)) in lines.iter().enumerate() { - if i >= inner.height as usize { break; } - f.render_widget(Paragraph::new(*text).style(*style), Rect::new(inner.x + 1, inner.y + i as u16, inner.width.saturating_sub(2), 1)); - } -} - -// ── Helpers ────────────────────────────────────────────────────────────────── - -fn mode_name(mode: &AppMode) -> &'static str { - match mode { + let mode_badge = match &app.mode { AppMode::Normal => "NORMAL", AppMode::Editing { .. } => "INSERT", AppMode::FormulaEdit { .. } => "FORMULA", @@ -442,27 +421,168 @@ fn mode_name(mode: &AppMode) -> &'static str { AppMode::CommandMode { .. } => "COMMAND", AppMode::Help => "HELP", AppMode::Quit => "QUIT", - } -} + }; -fn mode_style(mode: &AppMode) -> Style { - match mode { + let search_part = if app.search_mode { + format!(" /{}▌", app.search_query) + } else { + String::new() + }; + + let msg = if !app.status_msg.is_empty() { + app.status_msg.as_str() + } else { + app.hint_text() + }; + + let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; + let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); + + let left = format!(" {mode_badge}{search_part} {msg}"); + let right = view_badge; + let pad = " ".repeat((area.width as usize).saturating_sub(left.len() + right.len())); + let line = format!("{left}{pad}{right}"); + + let badge_style = match &app.mode { AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), _ => Style::default().fg(Color::Black).bg(Color::DarkGray), - } + }; + + f.render_widget(Paragraph::new(line).style(badge_style), area); } -fn fill_line(left: String, right: &str, width: u16) -> String { - let pad = " ".repeat((width as usize).saturating_sub(left.len() + right.len())); - format!("{left}{pad}{right}") +fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { + let buf = if let AppMode::CommandMode { buffer } = &app.mode { + buffer.as_str() + } else { + "" + }; + let line = format!(":{buf}▌"); + f.render_widget( + Paragraph::new(line).style(Style::default().fg(Color::White).bg(Color::Black)), + area, + ); } -fn centered_popup(area: Rect, width: u16, height: u16) -> Rect { - let w = width.min(area.width); - let h = height.min(area.height); +fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { + let buf = if let AppMode::ExportPrompt { buffer } = &app.mode { + buffer.as_str() + } else { + "" + }; + let popup_w = 64u16.min(area.width); + let x = area.x + area.width.saturating_sub(popup_w) / 2; + let y = area.y + area.height / 2; + let popup_area = Rect::new(x, y, popup_w, 3); + + f.render_widget(Clear, popup_area); + let block = Block::default() + .borders(Borders::ALL) + .border_style(Style::default().fg(Color::Yellow)) + .title(" Export CSV — path (Esc cancel) "); + let inner = block.inner(popup_area); + f.render_widget(block, popup_area); + f.render_widget( + Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), + inner, + ); +} + +fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { + let w = 58u16.min(area.width.saturating_sub(4)); + let h = 20u16.min(area.height.saturating_sub(2)); let x = area.x + area.width.saturating_sub(w) / 2; let y = area.y + area.height.saturating_sub(h) / 2; - Rect::new(x, y, w, h) + let popup = Rect::new(x, y, w, h); + + f.render_widget(Clear, popup); + + let block = Block::default() + .borders(Borders::ALL) + .border_style(Style::default().fg(Color::Blue)) + .title(" Welcome to improvise "); + let inner = block.inner(popup); + f.render_widget(block, popup); + + let lines: &[(&str, Style)] = &[ + ( + "Multi-dimensional data modeling — in your terminal.", + Style::default().fg(Color::White), + ), + ("", Style::default()), + ( + "Getting started", + Style::default() + .fg(Color::Blue) + .add_modifier(Modifier::BOLD), + ), + ("", Style::default()), + ( + ":import Import a JSON file", + Style::default().fg(Color::Cyan), + ), + ( + ":add-cat Add a category (dimension)", + Style::default().fg(Color::Cyan), + ), + ( + ":add-item Add an item to a category", + Style::default().fg(Color::Cyan), + ), + ( + ":formula Add a formula, e.g.:", + Style::default().fg(Color::Cyan), + ), + ( + " Profit = Revenue - Cost", + Style::default().fg(Color::Green), + ), + ( + ":w Save your model", + Style::default().fg(Color::Cyan), + ), + ("", Style::default()), + ( + "Navigation", + Style::default() + .fg(Color::Blue) + .add_modifier(Modifier::BOLD), + ), + ("", Style::default()), + ( + "F C V Open panels (Formulas/Categories/Views)", + Style::default(), + ), + ( + "T Tile-select: pivot rows ↔ cols ↔ page", + Style::default(), + ), + ("i Enter Edit a cell", Style::default()), + ( + "[ ] Cycle the page-axis filter", + Style::default(), + ), + ( + "? or :help Full key reference", + Style::default(), + ), + (":q Quit", Style::default()), + ]; + + for (i, (text, style)) in lines.iter().enumerate() { + if i >= inner.height as usize { + break; + } + f.render_widget( + Paragraph::new(*text).style(*style), + Rect::new( + inner.x + 1, + inner.y + i as u16, + inner.width.saturating_sub(2), + 1, + ), + ); + } } -- 2.51.0 From bbebc3344ceb641893e19de46ebb3367865470c4 Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:32:07 -0700 Subject: [PATCH 08/15] refactor: mystery model #1 --- src/main.rs | 54 +++++++++++++++++++++++++++-------------------------- 1 file changed, 28 insertions(+), 26 deletions(-) diff --git a/src/main.rs b/src/main.rs index 58526be..1d63685 100644 --- a/src/main.rs +++ b/src/main.rs @@ -466,24 +466,38 @@ fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { ); } +fn center_rect(area: Rect, width: u16, height: u16) -> Rect { + let w = width.min(area.width); + let h = height.min(area.height); + let x = area.x + area.width.saturating_sub(w) / 2; + let y = area.y + area.height.saturating_sub(h) / 2; + Rect::new(x, y, w, h) +} + +fn draw_popup_frame( + f: &mut Frame, + popup: Rect, + title: &str, + border_color: Color, +) -> Rect { + f.render_widget(Clear, popup); + let block = Block::default() + .borders(Borders::ALL) + .border_style(Style::default().fg(border_color)) + .title(title); + let inner = block.inner(popup); + f.render_widget(block, popup); + inner +} + fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { let buf = if let AppMode::ExportPrompt { buffer } = &app.mode { buffer.as_str() } else { "" }; - let popup_w = 64u16.min(area.width); - let x = area.x + area.width.saturating_sub(popup_w) / 2; - let y = area.y + area.height / 2; - let popup_area = Rect::new(x, y, popup_w, 3); - - f.render_widget(Clear, popup_area); - let block = Block::default() - .borders(Borders::ALL) - .border_style(Style::default().fg(Color::Yellow)) - .title(" Export CSV — path (Esc cancel) "); - let inner = block.inner(popup_area); - f.render_widget(block, popup_area); + let popup = center_rect(area, 64, 3); + let inner = draw_popup_frame(f, popup, " Export CSV — path (Esc cancel) ", Color::Yellow); f.render_widget( Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner, @@ -491,20 +505,8 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { } fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { - let w = 58u16.min(area.width.saturating_sub(4)); - let h = 20u16.min(area.height.saturating_sub(2)); - let x = area.x + area.width.saturating_sub(w) / 2; - let y = area.y + area.height.saturating_sub(h) / 2; - let popup = Rect::new(x, y, w, h); - - f.render_widget(Clear, popup); - - let block = Block::default() - .borders(Borders::ALL) - .border_style(Style::default().fg(Color::Blue)) - .title(" Welcome to improvise "); - let inner = block.inner(popup); - f.render_widget(block, popup); + let popup = center_rect(area, 58, 20); + let inner = draw_popup_frame(f, popup, " Welcome to improvise ", Color::Blue); let lines: &[(&str, Style)] = &[ ( -- 2.51.0 From a57d3ed294d317f6c9c89e3f2bc3e97e52c0b1eb Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:32:12 -0700 Subject: [PATCH 09/15] Revert "refactor: mystery model #1" This reverts commit bbebc3344ceb641893e19de46ebb3367865470c4. --- src/main.rs | 54 ++++++++++++++++++++++++++--------------------------- 1 file changed, 26 insertions(+), 28 deletions(-) diff --git a/src/main.rs b/src/main.rs index 1d63685..58526be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -466,38 +466,24 @@ fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { ); } -fn center_rect(area: Rect, width: u16, height: u16) -> Rect { - let w = width.min(area.width); - let h = height.min(area.height); - let x = area.x + area.width.saturating_sub(w) / 2; - let y = area.y + area.height.saturating_sub(h) / 2; - Rect::new(x, y, w, h) -} - -fn draw_popup_frame( - f: &mut Frame, - popup: Rect, - title: &str, - border_color: Color, -) -> Rect { - f.render_widget(Clear, popup); - let block = Block::default() - .borders(Borders::ALL) - .border_style(Style::default().fg(border_color)) - .title(title); - let inner = block.inner(popup); - f.render_widget(block, popup); - inner -} - fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { let buf = if let AppMode::ExportPrompt { buffer } = &app.mode { buffer.as_str() } else { "" }; - let popup = center_rect(area, 64, 3); - let inner = draw_popup_frame(f, popup, " Export CSV — path (Esc cancel) ", Color::Yellow); + let popup_w = 64u16.min(area.width); + let x = area.x + area.width.saturating_sub(popup_w) / 2; + let y = area.y + area.height / 2; + let popup_area = Rect::new(x, y, popup_w, 3); + + f.render_widget(Clear, popup_area); + let block = Block::default() + .borders(Borders::ALL) + .border_style(Style::default().fg(Color::Yellow)) + .title(" Export CSV — path (Esc cancel) "); + let inner = block.inner(popup_area); + f.render_widget(block, popup_area); f.render_widget( Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner, @@ -505,8 +491,20 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { } fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { - let popup = center_rect(area, 58, 20); - let inner = draw_popup_frame(f, popup, " Welcome to improvise ", Color::Blue); + let w = 58u16.min(area.width.saturating_sub(4)); + let h = 20u16.min(area.height.saturating_sub(2)); + let x = area.x + area.width.saturating_sub(w) / 2; + let y = area.y + area.height.saturating_sub(h) / 2; + let popup = Rect::new(x, y, w, h); + + f.render_widget(Clear, popup); + + let block = Block::default() + .borders(Borders::ALL) + .border_style(Style::default().fg(Color::Blue)) + .title(" Welcome to improvise "); + let inner = block.inner(popup); + f.render_widget(block, popup); let lines: &[(&str, Style)] = &[ ( -- 2.51.0 From 87fd6a16203f9f302ecf4381993832254a4e4739 Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:40:22 -0700 Subject: [PATCH 10/15] refactor: mystery model #2 --- src/main.rs | 168 ++++++++++++++++++++-------------------------------- 1 file changed, 64 insertions(+), 104 deletions(-) diff --git a/src/main.rs b/src/main.rs index 58526be..8e2f0cc 100644 --- a/src/main.rs +++ b/src/main.rs @@ -346,18 +346,13 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { fn draw_content(f: &mut Frame, area: Rect, app: &App) { let side_open = app.formula_panel_open || app.category_panel_open || app.view_panel_open; - if side_open { + let grid_area = if side_open { let side_w = 32u16; let chunks = Layout::default() .direction(Direction::Horizontal) .constraints([Constraint::Min(40), Constraint::Length(side_w)]) .split(area); - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - chunks[0], - ); - let side = chunks[1]; let panel_count = [ app.formula_panel_open, @@ -366,39 +361,36 @@ fn draw_content(f: &mut Frame, area: Rect, app: &App) { ] .iter() .filter(|&&b| b) - .count() as u16; - let ph = side.height / panel_count.max(1); - let mut y = side.y; + .count(); + + let constraints: Vec = std::iter::repeat(Constraint::Fill(1)) + .take(panel_count) + .collect(); + let panel_areas = Layout::default() + .direction(Direction::Vertical) + .constraints(constraints) + .split(side); + let mut slots = panel_areas.iter().copied(); if app.formula_panel_open { - let a = Rect::new(side.x, y, side.width, ph); - f.render_widget( - FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), - a, - ); - y += ph; + f.render_widget(FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), slots.next().unwrap()); } if app.category_panel_open { - let a = Rect::new(side.x, y, side.width, ph); - f.render_widget( - CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), - a, - ); - y += ph; + f.render_widget(CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), slots.next().unwrap()); } if app.view_panel_open { - let a = Rect::new(side.x, y, side.width, ph); - f.render_widget( - ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), - a, - ); + f.render_widget(ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), slots.next().unwrap()); } + + chunks[0] } else { - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - area, - ); - } + area + }; + + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + grid_area, + ); } fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { @@ -490,6 +482,37 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { ); } +#[derive(Copy, Clone)] +enum LineKind { + Intro, + Header, + Command, + Example, + Plain, +} + +const WELCOME_LINES: &[(&str, LineKind)] = &[ + ("Multi-dimensional data modeling — in your terminal.", LineKind::Intro), + ("", LineKind::Plain), + ("Getting started", LineKind::Header), + ("", LineKind::Plain), + (":import Import a JSON file", LineKind::Command), + (":add-cat Add a category (dimension)", LineKind::Command), + (":add-item Add an item to a category", LineKind::Command), + (":formula Add a formula, e.g.:", LineKind::Command), + (" Profit = Revenue - Cost", LineKind::Example), + (":w Save your model", LineKind::Command), + ("", LineKind::Plain), + ("Navigation", LineKind::Header), + ("", LineKind::Plain), + ("F C V Open panels (Formulas/Categories/Views)", LineKind::Plain), + ("T Tile-select: pivot rows ↔ cols ↔ page", LineKind::Plain), + ("i Enter Edit a cell", LineKind::Plain), + ("[ ] Cycle the page-axis filter", LineKind::Plain), + ("? or :help Full key reference", LineKind::Plain), + (":q Quit", LineKind::Plain), +]; + fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { let w = 58u16.min(area.width.saturating_sub(4)); let h = 20u16.min(area.height.saturating_sub(2)); @@ -506,83 +529,20 @@ fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { let inner = block.inner(popup); f.render_widget(block, popup); - let lines: &[(&str, Style)] = &[ - ( - "Multi-dimensional data modeling — in your terminal.", - Style::default().fg(Color::White), - ), - ("", Style::default()), - ( - "Getting started", - Style::default() - .fg(Color::Blue) - .add_modifier(Modifier::BOLD), - ), - ("", Style::default()), - ( - ":import Import a JSON file", - Style::default().fg(Color::Cyan), - ), - ( - ":add-cat Add a category (dimension)", - Style::default().fg(Color::Cyan), - ), - ( - ":add-item Add an item to a category", - Style::default().fg(Color::Cyan), - ), - ( - ":formula Add a formula, e.g.:", - Style::default().fg(Color::Cyan), - ), - ( - " Profit = Revenue - Cost", - Style::default().fg(Color::Green), - ), - ( - ":w Save your model", - Style::default().fg(Color::Cyan), - ), - ("", Style::default()), - ( - "Navigation", - Style::default() - .fg(Color::Blue) - .add_modifier(Modifier::BOLD), - ), - ("", Style::default()), - ( - "F C V Open panels (Formulas/Categories/Views)", - Style::default(), - ), - ( - "T Tile-select: pivot rows ↔ cols ↔ page", - Style::default(), - ), - ("i Enter Edit a cell", Style::default()), - ( - "[ ] Cycle the page-axis filter", - Style::default(), - ), - ( - "? or :help Full key reference", - Style::default(), - ), - (":q Quit", Style::default()), - ]; - - for (i, (text, style)) in lines.iter().enumerate() { + for (i, (text, kind)) in WELCOME_LINES.iter().enumerate() { if i >= inner.height as usize { break; } + let style = match kind { + LineKind::Intro => Style::default().fg(Color::White), + LineKind::Header => Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD), + LineKind::Command => Style::default().fg(Color::Cyan), + LineKind::Example => Style::default().fg(Color::Green), + LineKind::Plain => Style::default(), + }; f.render_widget( - Paragraph::new(*text).style(*style), - Rect::new( - inner.x + 1, - inner.y + i as u16, - inner.width.saturating_sub(2), - 1, - ), + Paragraph::new(*text).style(style), + Rect::new(inner.x + 1, inner.y + i as u16, inner.width.saturating_sub(2), 1), ); } } -- 2.51.0 From 6d88de302096b2049800651721040bfafb53c43f Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:41:25 -0700 Subject: [PATCH 11/15] Revert "refactor: mystery model #2" This reverts commit 87fd6a16203f9f302ecf4381993832254a4e4739. --- src/main.rs | 168 ++++++++++++++++++++++++++++++++-------------------- 1 file changed, 104 insertions(+), 64 deletions(-) diff --git a/src/main.rs b/src/main.rs index 8e2f0cc..58526be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -346,13 +346,18 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { fn draw_content(f: &mut Frame, area: Rect, app: &App) { let side_open = app.formula_panel_open || app.category_panel_open || app.view_panel_open; - let grid_area = if side_open { + if side_open { let side_w = 32u16; let chunks = Layout::default() .direction(Direction::Horizontal) .constraints([Constraint::Min(40), Constraint::Length(side_w)]) .split(area); + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + chunks[0], + ); + let side = chunks[1]; let panel_count = [ app.formula_panel_open, @@ -361,36 +366,39 @@ fn draw_content(f: &mut Frame, area: Rect, app: &App) { ] .iter() .filter(|&&b| b) - .count(); - - let constraints: Vec = std::iter::repeat(Constraint::Fill(1)) - .take(panel_count) - .collect(); - let panel_areas = Layout::default() - .direction(Direction::Vertical) - .constraints(constraints) - .split(side); - let mut slots = panel_areas.iter().copied(); + .count() as u16; + let ph = side.height / panel_count.max(1); + let mut y = side.y; if app.formula_panel_open { - f.render_widget(FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), slots.next().unwrap()); + let a = Rect::new(side.x, y, side.width, ph); + f.render_widget( + FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), + a, + ); + y += ph; } if app.category_panel_open { - f.render_widget(CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), slots.next().unwrap()); + let a = Rect::new(side.x, y, side.width, ph); + f.render_widget( + CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), + a, + ); + y += ph; } if app.view_panel_open { - f.render_widget(ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), slots.next().unwrap()); + let a = Rect::new(side.x, y, side.width, ph); + f.render_widget( + ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), + a, + ); } - - chunks[0] } else { - area - }; - - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - grid_area, - ); + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + area, + ); + } } fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { @@ -482,37 +490,6 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { ); } -#[derive(Copy, Clone)] -enum LineKind { - Intro, - Header, - Command, - Example, - Plain, -} - -const WELCOME_LINES: &[(&str, LineKind)] = &[ - ("Multi-dimensional data modeling — in your terminal.", LineKind::Intro), - ("", LineKind::Plain), - ("Getting started", LineKind::Header), - ("", LineKind::Plain), - (":import Import a JSON file", LineKind::Command), - (":add-cat Add a category (dimension)", LineKind::Command), - (":add-item Add an item to a category", LineKind::Command), - (":formula Add a formula, e.g.:", LineKind::Command), - (" Profit = Revenue - Cost", LineKind::Example), - (":w Save your model", LineKind::Command), - ("", LineKind::Plain), - ("Navigation", LineKind::Header), - ("", LineKind::Plain), - ("F C V Open panels (Formulas/Categories/Views)", LineKind::Plain), - ("T Tile-select: pivot rows ↔ cols ↔ page", LineKind::Plain), - ("i Enter Edit a cell", LineKind::Plain), - ("[ ] Cycle the page-axis filter", LineKind::Plain), - ("? or :help Full key reference", LineKind::Plain), - (":q Quit", LineKind::Plain), -]; - fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { let w = 58u16.min(area.width.saturating_sub(4)); let h = 20u16.min(area.height.saturating_sub(2)); @@ -529,20 +506,83 @@ fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { let inner = block.inner(popup); f.render_widget(block, popup); - for (i, (text, kind)) in WELCOME_LINES.iter().enumerate() { + let lines: &[(&str, Style)] = &[ + ( + "Multi-dimensional data modeling — in your terminal.", + Style::default().fg(Color::White), + ), + ("", Style::default()), + ( + "Getting started", + Style::default() + .fg(Color::Blue) + .add_modifier(Modifier::BOLD), + ), + ("", Style::default()), + ( + ":import Import a JSON file", + Style::default().fg(Color::Cyan), + ), + ( + ":add-cat Add a category (dimension)", + Style::default().fg(Color::Cyan), + ), + ( + ":add-item Add an item to a category", + Style::default().fg(Color::Cyan), + ), + ( + ":formula Add a formula, e.g.:", + Style::default().fg(Color::Cyan), + ), + ( + " Profit = Revenue - Cost", + Style::default().fg(Color::Green), + ), + ( + ":w Save your model", + Style::default().fg(Color::Cyan), + ), + ("", Style::default()), + ( + "Navigation", + Style::default() + .fg(Color::Blue) + .add_modifier(Modifier::BOLD), + ), + ("", Style::default()), + ( + "F C V Open panels (Formulas/Categories/Views)", + Style::default(), + ), + ( + "T Tile-select: pivot rows ↔ cols ↔ page", + Style::default(), + ), + ("i Enter Edit a cell", Style::default()), + ( + "[ ] Cycle the page-axis filter", + Style::default(), + ), + ( + "? or :help Full key reference", + Style::default(), + ), + (":q Quit", Style::default()), + ]; + + for (i, (text, style)) in lines.iter().enumerate() { if i >= inner.height as usize { break; } - let style = match kind { - LineKind::Intro => Style::default().fg(Color::White), - LineKind::Header => Style::default().fg(Color::Blue).add_modifier(Modifier::BOLD), - LineKind::Command => Style::default().fg(Color::Cyan), - LineKind::Example => Style::default().fg(Color::Green), - LineKind::Plain => Style::default(), - }; f.render_widget( - Paragraph::new(*text).style(style), - Rect::new(inner.x + 1, inner.y + i as u16, inner.width.saturating_sub(2), 1), + Paragraph::new(*text).style(*style), + Rect::new( + inner.x + 1, + inner.y + i as u16, + inner.width.saturating_sub(2), + 1, + ), ); } } -- 2.51.0 From 4b721f754384bab4603ff901094cd27aeb0ae22c Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:46:25 -0700 Subject: [PATCH 12/15] refactor: mystery model 3 --- src/main.rs | 66 ++++++++++++++++++++++++++--------------------------- 1 file changed, 32 insertions(+), 34 deletions(-) diff --git a/src/main.rs b/src/main.rs index 58526be..ddb3097 100644 --- a/src/main.rs +++ b/src/main.rs @@ -344,61 +344,59 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { } fn draw_content(f: &mut Frame, area: Rect, app: &App) { - let side_open = app.formula_panel_open || app.category_panel_open || app.view_panel_open; + let panel_count = [ + app.formula_panel_open, + app.category_panel_open, + app.view_panel_open, + ] + .iter() + .filter(|&&b| b) + .count(); - if side_open { - let side_w = 32u16; + let grid_area = if panel_count == 0 { + area + } else { let chunks = Layout::default() .direction(Direction::Horizontal) - .constraints([Constraint::Min(40), Constraint::Length(side_w)]) + .constraints([Constraint::Min(40), Constraint::Length(32)]) .split(area); - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - chunks[0], - ); - - let side = chunks[1]; - let panel_count = [ - app.formula_panel_open, - app.category_panel_open, - app.view_panel_open, - ] - .iter() - .filter(|&&b| b) - .count() as u16; - let ph = side.height / panel_count.max(1); - let mut y = side.y; + let constraints: Vec = + vec![Constraint::Ratio(1, panel_count as u32); panel_count]; + let panel_areas = Layout::default() + .direction(Direction::Vertical) + .constraints(constraints) + .split(chunks[1]); + let mut slot = 0; if app.formula_panel_open { - let a = Rect::new(side.x, y, side.width, ph); f.render_widget( FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), - a, + panel_areas[slot], ); - y += ph; + slot += 1; } if app.category_panel_open { - let a = Rect::new(side.x, y, side.width, ph); f.render_widget( CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), - a, + panel_areas[slot], ); - y += ph; + slot += 1; } if app.view_panel_open { - let a = Rect::new(side.x, y, side.width, ph); f.render_widget( ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), - a, + panel_areas[slot], ); } - } else { - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - area, - ); - } + + chunks[0] + }; + + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + grid_area, + ); } fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { -- 2.51.0 From 3f84ba03cbd7d4308a221ad0524abd56780db31b Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 00:46:55 -0700 Subject: [PATCH 13/15] Revert "refactor: mystery model 3" This reverts commit 4b721f754384bab4603ff901094cd27aeb0ae22c. --- src/main.rs | 66 +++++++++++++++++++++++++++-------------------------- 1 file changed, 34 insertions(+), 32 deletions(-) diff --git a/src/main.rs b/src/main.rs index ddb3097..58526be 100644 --- a/src/main.rs +++ b/src/main.rs @@ -344,59 +344,61 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { } fn draw_content(f: &mut Frame, area: Rect, app: &App) { - let panel_count = [ - app.formula_panel_open, - app.category_panel_open, - app.view_panel_open, - ] - .iter() - .filter(|&&b| b) - .count(); + let side_open = app.formula_panel_open || app.category_panel_open || app.view_panel_open; - let grid_area = if panel_count == 0 { - area - } else { + if side_open { + let side_w = 32u16; let chunks = Layout::default() .direction(Direction::Horizontal) - .constraints([Constraint::Min(40), Constraint::Length(32)]) + .constraints([Constraint::Min(40), Constraint::Length(side_w)]) .split(area); - let constraints: Vec = - vec![Constraint::Ratio(1, panel_count as u32); panel_count]; - let panel_areas = Layout::default() - .direction(Direction::Vertical) - .constraints(constraints) - .split(chunks[1]); + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + chunks[0], + ); + + let side = chunks[1]; + let panel_count = [ + app.formula_panel_open, + app.category_panel_open, + app.view_panel_open, + ] + .iter() + .filter(|&&b| b) + .count() as u16; + let ph = side.height / panel_count.max(1); + let mut y = side.y; - let mut slot = 0; if app.formula_panel_open { + let a = Rect::new(side.x, y, side.width, ph); f.render_widget( FormulaPanel::new(&app.model, &app.mode, app.formula_cursor), - panel_areas[slot], + a, ); - slot += 1; + y += ph; } if app.category_panel_open { + let a = Rect::new(side.x, y, side.width, ph); f.render_widget( CategoryPanel::new(&app.model, &app.mode, app.cat_panel_cursor), - panel_areas[slot], + a, ); - slot += 1; + y += ph; } if app.view_panel_open { + let a = Rect::new(side.x, y, side.width, ph); f.render_widget( ViewPanel::new(&app.model, &app.mode, app.view_panel_cursor), - panel_areas[slot], + a, ); } - - chunks[0] - }; - - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - grid_area, - ); + } else { + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + area, + ); + } } fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { -- 2.51.0 From 9fc3f0b5d6b6fa25d31596b071834d021940f065 Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 01:01:19 -0700 Subject: [PATCH 14/15] refactor: synthesize previous refactors --- src/main.rs | 164 ++++++++++++++++++++++++++-------------------------- 1 file changed, 81 insertions(+), 83 deletions(-) diff --git a/src/main.rs b/src/main.rs index 58526be..edb9ed6 100644 --- a/src/main.rs +++ b/src/main.rs @@ -277,11 +277,61 @@ fn run_tui( // ── Drawing ────────────────────────────────────────────────────────────────── +fn fill_line(left: String, right: &str, width: u16) -> String { + let pad = " ".repeat((width as usize).saturating_sub(left.len() + right.len())); + format!("{left}{pad}{right}") +} + +fn centered_popup(area: Rect, width: u16, height: u16) -> Rect { + let w = width.min(area.width); + let h = height.min(area.height); + let x = area.x + area.width.saturating_sub(w) / 2; + let y = area.y + area.height.saturating_sub(h) / 2; + Rect::new(x, y, w, h) +} + +fn draw_popup_frame(f: &mut Frame, popup: Rect, title: &str, border_color: Color) -> Rect { + f.render_widget(Clear, popup); + let block = Block::default() + .borders(Borders::ALL) + .border_style(Style::default().fg(border_color)) + .title(title); + let inner = block.inner(popup); + f.render_widget(block, popup); + inner +} + +fn mode_name(mode: &AppMode) -> &'static str { + match mode { + AppMode::Normal => "NORMAL", + AppMode::Editing { .. } => "INSERT", + AppMode::FormulaEdit { .. } => "FORMULA", + AppMode::FormulaPanel => "FORMULAS", + AppMode::CategoryPanel => "CATEGORIES", + AppMode::CategoryAdd { .. } => "NEW CATEGORY", + AppMode::ItemAdd { .. } => "ADD ITEMS", + AppMode::ViewPanel => "VIEWS", + AppMode::TileSelect { .. } => "TILES", + AppMode::ImportWizard => "IMPORT", + AppMode::ExportPrompt { .. } => "EXPORT", + AppMode::CommandMode { .. } => "COMMAND", + AppMode::Help => "HELP", + AppMode::Quit => "QUIT", + } +} + +fn mode_style(mode: &AppMode) -> Style { + match mode { + AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), + AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), + AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), + _ => Style::default().fg(Color::Black).bg(Color::DarkGray), + } +} + fn draw(f: &mut Frame, app: &App) { let size = f.area(); - let is_cmd_mode = matches!(app.mode, AppMode::CommandMode { .. }); - let main_chunks = Layout::default() .direction(Direction::Vertical) .constraints([ @@ -295,12 +345,7 @@ fn draw(f: &mut Frame, app: &App) { draw_title(f, main_chunks[0], app); draw_content(f, main_chunks[1], app); draw_tile_bar(f, main_chunks[2], app); - - if is_cmd_mode { - draw_command_bar(f, main_chunks[3], app); - } else { - draw_status(f, main_chunks[3], app); - } + draw_bottom_bar(f, main_chunks[3], app); // Overlays (rendered last so they appear on top) if matches!(app.mode, AppMode::Help) { @@ -315,7 +360,7 @@ fn draw(f: &mut Frame, app: &App) { draw_export_prompt(f, size, app); } if app.is_empty_model() && matches!(app.mode, AppMode::Normal | AppMode::CommandMode { .. }) { - draw_welcome(f, main_chunks[1], app); + draw_welcome(f, main_chunks[1]); } } @@ -330,8 +375,7 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { .unwrap_or_default(); let title = format!(" improvise · {}{}{} ", app.model.name, file, dirty); let right = " ?:help :q quit "; - let pad = " ".repeat((area.width as usize).saturating_sub(title.len() + right.len())); - let line = format!("{title}{pad}{right}"); + let line = fill_line(title, right, area.width); f.render_widget( Paragraph::new(line).style( Style::default() @@ -346,6 +390,7 @@ fn draw_title(f: &mut Frame, area: Rect, app: &App) { fn draw_content(f: &mut Frame, area: Rect, app: &App) { let side_open = app.formula_panel_open || app.category_panel_open || app.view_panel_open; + let grid_area; if side_open { let side_w = 32u16; let chunks = Layout::default() @@ -353,10 +398,7 @@ fn draw_content(f: &mut Frame, area: Rect, app: &App) { .constraints([Constraint::Min(40), Constraint::Length(side_w)]) .split(area); - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - chunks[0], - ); + grid_area = chunks[0]; let side = chunks[1]; let panel_count = [ @@ -394,35 +436,27 @@ fn draw_content(f: &mut Frame, area: Rect, app: &App) { ); } } else { - f.render_widget( - GridWidget::new(&app.model, &app.mode, &app.search_query), - area, - ); + grid_area = area; } + + f.render_widget( + GridWidget::new(&app.model, &app.mode, &app.search_query), + grid_area, + ); } fn draw_tile_bar(f: &mut Frame, area: Rect, app: &App) { f.render_widget(TileBar::new(&app.model, &app.mode), area); } -fn draw_status(f: &mut Frame, area: Rect, app: &App) { - let mode_badge = match &app.mode { - AppMode::Normal => "NORMAL", - AppMode::Editing { .. } => "INSERT", - AppMode::FormulaEdit { .. } => "FORMULA", - AppMode::FormulaPanel => "FORMULAS", - AppMode::CategoryPanel => "CATEGORIES", - AppMode::CategoryAdd { .. } => "NEW CATEGORY", - AppMode::ItemAdd { .. } => "ADD ITEMS", - AppMode::ViewPanel => "VIEWS", - AppMode::TileSelect { .. } => "TILES", - AppMode::ImportWizard => "IMPORT", - AppMode::ExportPrompt { .. } => "EXPORT", - AppMode::CommandMode { .. } => "COMMAND", - AppMode::Help => "HELP", - AppMode::Quit => "QUIT", - }; +fn draw_bottom_bar(f: &mut Frame, area: Rect, app: &App) { + match app.mode { + AppMode::CommandMode { ref buffer } => draw_command_bar(f, area, buffer), + _ => draw_status(f, area, app), + } +} +fn draw_status(f: &mut Frame, area: Rect, app: &App) { let search_part = if app.search_mode { format!(" /{}▌", app.search_query) } else { @@ -438,30 +472,16 @@ fn draw_status(f: &mut Frame, area: Rect, app: &App) { let yank_indicator = if app.yanked.is_some() { " [yank]" } else { "" }; let view_badge = format!(" {}{} ", app.model.active_view, yank_indicator); - let left = format!(" {mode_badge}{search_part} {msg}"); - let right = view_badge; - let pad = " ".repeat((area.width as usize).saturating_sub(left.len() + right.len())); - let line = format!("{left}{pad}{right}"); + let left = format!(" {}{search_part} {msg}", mode_name(&app.mode)); + let line = fill_line(left, &view_badge, area.width); - let badge_style = match &app.mode { - AppMode::Editing { .. } => Style::default().fg(Color::Black).bg(Color::Green), - AppMode::CommandMode { .. } => Style::default().fg(Color::Black).bg(Color::Yellow), - AppMode::TileSelect { .. } => Style::default().fg(Color::Black).bg(Color::Magenta), - _ => Style::default().fg(Color::Black).bg(Color::DarkGray), - }; - - f.render_widget(Paragraph::new(line).style(badge_style), area); + f.render_widget(Paragraph::new(line).style(mode_style(&app.mode)), area); } -fn draw_command_bar(f: &mut Frame, area: Rect, app: &App) { - let buf = if let AppMode::CommandMode { buffer } = &app.mode { - buffer.as_str() - } else { - "" - }; - let line = format!(":{buf}▌"); +fn draw_command_bar(f: &mut Frame, area: Rect, buffer: &str) { f.render_widget( - Paragraph::new(line).style(Style::default().fg(Color::White).bg(Color::Black)), + Paragraph::new(format!(":{buffer}▌")) + .style(Style::default().fg(Color::White).bg(Color::Black)), area, ); } @@ -472,39 +492,17 @@ fn draw_export_prompt(f: &mut Frame, area: Rect, app: &App) { } else { "" }; - let popup_w = 64u16.min(area.width); - let x = area.x + area.width.saturating_sub(popup_w) / 2; - let y = area.y + area.height / 2; - let popup_area = Rect::new(x, y, popup_w, 3); - - f.render_widget(Clear, popup_area); - let block = Block::default() - .borders(Borders::ALL) - .border_style(Style::default().fg(Color::Yellow)) - .title(" Export CSV — path (Esc cancel) "); - let inner = block.inner(popup_area); - f.render_widget(block, popup_area); + let popup = centered_popup(area, 64, 3); + let inner = draw_popup_frame(f, popup, " Export CSV — path (Esc cancel) ", Color::Yellow); f.render_widget( Paragraph::new(format!("{buf}▌")).style(Style::default().fg(Color::Green)), inner, ); } -fn draw_welcome(f: &mut Frame, area: Rect, _app: &App) { - let w = 58u16.min(area.width.saturating_sub(4)); - let h = 20u16.min(area.height.saturating_sub(2)); - let x = area.x + area.width.saturating_sub(w) / 2; - let y = area.y + area.height.saturating_sub(h) / 2; - let popup = Rect::new(x, y, w, h); - - f.render_widget(Clear, popup); - - let block = Block::default() - .borders(Borders::ALL) - .border_style(Style::default().fg(Color::Blue)) - .title(" Welcome to improvise "); - let inner = block.inner(popup); - f.render_widget(block, popup); +fn draw_welcome(f: &mut Frame, area: Rect) { + let popup = centered_popup(area, 58, 20); + let inner = draw_popup_frame(f, popup, " Welcome to improvise ", Color::Blue); let lines: &[(&str, Style)] = &[ ( -- 2.51.0 From ef79a3972196b735c3be426f4deb48fb658756a8 Mon Sep 17 00:00:00 2001 From: Edward Langley Date: Wed, 1 Apr 2026 01:32:19 -0700 Subject: [PATCH 15/15] Add CSV import functionality - Use csv crate for robust CSV parsing (handles quoted fields, empty values, \r\n) - Extend --import command to auto-detect format by file extension (.csv or .json) - Reuse existing ImportPipeline and analyzer for field type detection - Categories detected automatically (string fields), measures for numeric fields - Updated help text and welcome screen to mention CSV support All 201 tests pass. --- Cargo.lock | 23 + Cargo.toml | 2 + llama-server.log | 12789 +++++++++++++++++++++++++++++++++++++ src/command/dispatch.rs | 69 +- src/import/csv_parser.rs | 159 + src/import/mod.rs | 1 + src/main.rs | 40 +- 7 files changed, 13045 insertions(+), 38 deletions(-) create mode 100644 llama-server.log create mode 100644 src/import/csv_parser.rs diff --git a/Cargo.lock b/Cargo.lock index e40d413..3a325be 100644 --- a/Cargo.lock +++ b/Cargo.lock @@ -161,6 +161,27 @@ dependencies = [ "winapi", ] +[[package]] +name = "csv" +version = "1.4.0" +source = "registry+https://github.com/rust-lang/crates.io-index" +checksum = "52cd9d68cf7efc6ddfaaee42e7288d3a99d613d4b50f76ce9827ae0c6e14f938" +dependencies = [ + "csv-core", + "itoa", + "ryu", + "serde_core", +] + +[[package]] +name = "csv-core" +version = "0.1.13" +source = "registry+https://github.com/rust-lang/crates.io-index" +checksum = "704a3c26996a80471189265814dbc2c257598b96b8a7feae2d31ace646bb9782" +dependencies = [ + "memchr", +] + [[package]] name = "darling" version = "0.23.0" @@ -374,6 +395,7 @@ dependencies = [ "anyhow", "chrono", "crossterm", + "csv", "dirs", "flate2", "indexmap", @@ -381,6 +403,7 @@ dependencies = [ "ratatui", "serde", "serde_json", + "tempfile", "thiserror", "unicode-width 0.2.0", ] diff --git a/Cargo.toml b/Cargo.toml index 01e4281..993994e 100644 --- a/Cargo.toml +++ b/Cargo.toml @@ -21,9 +21,11 @@ chrono = { version = "0.4", features = ["serde"] } flate2 = "1" unicode-width = "0.2" dirs = "5" +csv = "1" [dev-dependencies] proptest = "1" +tempfile = "3" [profile.release] opt-level = 3 diff --git a/llama-server.log b/llama-server.log new file mode 100644 index 0000000..3d21b4f --- /dev/null +++ b/llama-server.log @@ -0,0 +1,12789 @@ +ggml_cuda_init: found 1 ROCm devices (Total VRAM: 32752 MiB): + Device 0: AMD Instinct MI100, gfx908:sramecc+:xnack- (0x908), VMM: no, Wave Size: 64, VRAM: 32752 MiB +common_download_file_single_online: no previous model file found /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_preset.ini +common_download_file_single_online: HEAD failed, status: 404 +no remote preset found, skipping +common_download_file_single_online: using cached file (same etag): /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf +common_download_file_single_online: using cached file (same etag): /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00002-of-00003.gguf +common_download_file_single_online: using cached file (same etag): /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00003-of-00003.gguf +build: 8414 (5744d7ec4) with GNU 14.3.0 for Linux x86_64 +system info: n_threads = 16, n_threads_batch = 16, total_threads = 16 + +system_info: n_threads = 16 (n_threads_batch = 16) / 16 | ROCm : NO_VMM = 1 | PEER_MAX_BATCH_SIZE = 128 | CPU : SSE3 = 1 | SSSE3 = 1 | AVX = 1 | AVX_VNNI = 1 | AVX2 = 1 | F16C = 1 | FMA = 1 | BMI2 = 1 | AVX512 = 1 | AVX512_VBMI = 1 | AVX512_VNNI = 1 | AVX512_BF16 = 1 | LLAMAFILE = 1 | OPENMP = 1 | REPACK = 1 | + +Running without SSL +init: using 15 threads for HTTP server +start: binding port with default address family +main: loading model +srv load_model: loading model '/home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf' +common_init_result: fitting params to device memory, for bugs during this step try to reproduce them with -fit off, or provide --verbose logs if the bug only occurs with -fit on +llama_params_fit_impl: getting device memory data for initial parameters: +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32734 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +create_tensor: loading tensor blk.23.ffn_down_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +create_tensor: loading tensor blk.24.ffn_down_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +create_tensor: loading tensor blk.25.ffn_down_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +create_tensor: loading tensor blk.26.ffn_down_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +create_tensor: loading tensor blk.27.ffn_down_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +create_tensor: loading tensor blk.28.ffn_down_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +create_tensor: loading tensor blk.29.ffn_down_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +create_tensor: loading tensor blk.30.ffn_down_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +create_tensor: loading tensor blk.31.ffn_down_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +create_tensor: loading tensor blk.32.ffn_down_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +create_tensor: loading tensor blk.33.ffn_down_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +create_tensor: loading tensor blk.34.ffn_down_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +create_tensor: loading tensor blk.35.ffn_down_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +create_tensor: loading tensor blk.36.ffn_down_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +create_tensor: loading tensor blk.37.ffn_down_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +create_tensor: loading tensor blk.38.ffn_down_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +create_tensor: loading tensor blk.39.ffn_down_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +create_tensor: loading tensor blk.40.ffn_down_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +create_tensor: loading tensor blk.41.ffn_down_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +create_tensor: loading tensor blk.42.ffn_down_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +create_tensor: loading tensor blk.43.ffn_down_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +create_tensor: loading tensor blk.44.ffn_down_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +create_tensor: loading tensor blk.45.ffn_down_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +create_tensor: loading tensor blk.46.ffn_down_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +create_tensor: loading tensor blk.47.ffn_down_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 0 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 420.01 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 2 +sched_reserve: reserve took 9.38 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + (57572 = 54004 + 3147 + 420) + 17592185987085 | +llama_memory_breakdown_print: | - Host | 468 = 204 + 0 + 264 | +llama_params_fit_impl: projected to use 57572 MiB of device memory vs. 32510 MiB of free device memory +llama_params_fit_impl: cannot meet free memory target of 1024 MiB, need to reduce device memory by 26086 MiB +llama_params_fit_impl: context size set by user to 131072 -> no change +llama_params_fit_impl: getting device memory data with all MoE tensors moved to system memory: +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +tensor blk.0.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.0.ffn_down_exps.weight +tensor blk.0.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +tensor blk.0.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +tensor blk.1.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.1.ffn_down_exps.weight +tensor blk.1.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +tensor blk.1.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +tensor blk.2.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.2.ffn_down_exps.weight +tensor blk.2.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +tensor blk.2.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +tensor blk.3.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.3.ffn_down_exps.weight +tensor blk.3.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +tensor blk.3.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +tensor blk.4.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.4.ffn_down_exps.weight +tensor blk.4.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +tensor blk.4.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +tensor blk.5.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.5.ffn_down_exps.weight +tensor blk.5.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +tensor blk.5.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +tensor blk.6.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.6.ffn_down_exps.weight +tensor blk.6.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +tensor blk.6.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +tensor blk.7.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.7.ffn_down_exps.weight +tensor blk.7.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +tensor blk.7.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +tensor blk.8.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.8.ffn_down_exps.weight +tensor blk.8.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +tensor blk.8.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +tensor blk.9.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.9.ffn_down_exps.weight +tensor blk.9.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +tensor blk.9.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +tensor blk.10.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.10.ffn_down_exps.weight +tensor blk.10.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +tensor blk.10.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +tensor blk.11.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.11.ffn_down_exps.weight +tensor blk.11.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +tensor blk.11.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +tensor blk.12.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.12.ffn_down_exps.weight +tensor blk.12.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +tensor blk.12.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +tensor blk.13.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.13.ffn_down_exps.weight +tensor blk.13.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +tensor blk.13.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +tensor blk.14.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.14.ffn_down_exps.weight +tensor blk.14.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +tensor blk.14.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +tensor blk.15.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.15.ffn_down_exps.weight +tensor blk.15.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +tensor blk.15.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +tensor blk.16.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.16.ffn_down_exps.weight +tensor blk.16.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +tensor blk.16.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +tensor blk.17.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.17.ffn_down_exps.weight +tensor blk.17.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +tensor blk.17.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +tensor blk.18.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.18.ffn_down_exps.weight +tensor blk.18.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +tensor blk.18.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +tensor blk.19.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.19.ffn_down_exps.weight +tensor blk.19.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +tensor blk.19.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +tensor blk.20.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.20.ffn_down_exps.weight +tensor blk.20.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +tensor blk.20.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +tensor blk.21.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.21.ffn_down_exps.weight +tensor blk.21.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +tensor blk.21.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +tensor blk.22.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.22.ffn_down_exps.weight +tensor blk.22.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +tensor blk.22.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +tensor blk.23.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_exps.weight +tensor blk.23.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +tensor blk.23.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +tensor blk.24.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_down_exps.weight +tensor blk.24.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +tensor blk.24.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +tensor blk.25.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_down_exps.weight +tensor blk.25.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +tensor blk.25.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +tensor blk.26.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_down_exps.weight +tensor blk.26.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +tensor blk.26.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +tensor blk.27.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_down_exps.weight +tensor blk.27.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +tensor blk.27.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +tensor blk.28.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_down_exps.weight +tensor blk.28.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +tensor blk.28.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +tensor blk.29.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_down_exps.weight +tensor blk.29.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +tensor blk.29.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +tensor blk.30.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_down_exps.weight +tensor blk.30.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +tensor blk.30.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +tensor blk.31.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_down_exps.weight +tensor blk.31.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +tensor blk.31.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +tensor blk.32.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_down_exps.weight +tensor blk.32.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +tensor blk.32.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +tensor blk.33.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_down_exps.weight +tensor blk.33.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +tensor blk.33.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +tensor blk.34.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_down_exps.weight +tensor blk.34.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +tensor blk.34.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +tensor blk.35.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_down_exps.weight +tensor blk.35.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +tensor blk.35.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +tensor blk.36.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_down_exps.weight +tensor blk.36.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +tensor blk.36.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +tensor blk.37.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_down_exps.weight +tensor blk.37.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +tensor blk.37.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +tensor blk.38.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_down_exps.weight +tensor blk.38.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +tensor blk.38.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +tensor blk.39.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_down_exps.weight +tensor blk.39.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +tensor blk.39.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +tensor blk.40.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_down_exps.weight +tensor blk.40.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +tensor blk.40.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +tensor blk.41.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_down_exps.weight +tensor blk.41.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +tensor blk.41.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +tensor blk.42.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_down_exps.weight +tensor blk.42.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +tensor blk.42.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +tensor blk.43.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_down_exps.weight +tensor blk.43.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +tensor blk.43.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +tensor blk.44.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_down_exps.weight +tensor blk.44.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +tensor blk.44.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +tensor blk.45.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_down_exps.weight +tensor blk.45.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +tensor blk.45.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +tensor blk.46.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_down_exps.weight +tensor blk.46.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +tensor blk.46.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +tensor blk.47.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_down_exps.weight +tensor blk.47.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +tensor blk.47.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 144 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +load_tensors: ROCm_Host model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 736.00 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 146 (with bs=512), 98 (with bs=1) +sched_reserve: reserve took 8.36 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + ( 5568 = 1684 + 3147 + 736) + 17592186039089 | +llama_memory_breakdown_print: | - Host | 52788 = 52524 + 0 + 264 | +llama_params_fit_impl: with only dense weights in device memory there is a total surplus of 25917 MiB +llama_params_fit_impl: id=0, target=31486 MiB +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device CPU, is_swa = 0 +load_tensors: layer 1 assigned to device CPU, is_swa = 0 +load_tensors: layer 2 assigned to device CPU, is_swa = 0 +load_tensors: layer 3 assigned to device CPU, is_swa = 0 +load_tensors: layer 4 assigned to device CPU, is_swa = 0 +load_tensors: layer 5 assigned to device CPU, is_swa = 0 +load_tensors: layer 6 assigned to device CPU, is_swa = 0 +load_tensors: layer 7 assigned to device CPU, is_swa = 0 +load_tensors: layer 8 assigned to device CPU, is_swa = 0 +load_tensors: layer 9 assigned to device CPU, is_swa = 0 +load_tensors: layer 10 assigned to device CPU, is_swa = 0 +load_tensors: layer 11 assigned to device CPU, is_swa = 0 +load_tensors: layer 12 assigned to device CPU, is_swa = 0 +load_tensors: layer 13 assigned to device CPU, is_swa = 0 +load_tensors: layer 14 assigned to device CPU, is_swa = 0 +load_tensors: layer 15 assigned to device CPU, is_swa = 0 +load_tensors: layer 16 assigned to device CPU, is_swa = 0 +load_tensors: layer 17 assigned to device CPU, is_swa = 0 +load_tensors: layer 18 assigned to device CPU, is_swa = 0 +load_tensors: layer 19 assigned to device CPU, is_swa = 0 +load_tensors: layer 20 assigned to device CPU, is_swa = 0 +load_tensors: layer 21 assigned to device CPU, is_swa = 0 +load_tensors: layer 22 assigned to device CPU, is_swa = 0 +load_tensors: layer 23 assigned to device CPU, is_swa = 0 +load_tensors: layer 24 assigned to device CPU, is_swa = 0 +load_tensors: layer 25 assigned to device CPU, is_swa = 0 +load_tensors: layer 26 assigned to device CPU, is_swa = 0 +load_tensors: layer 27 assigned to device CPU, is_swa = 0 +load_tensors: layer 28 assigned to device CPU, is_swa = 0 +load_tensors: layer 29 assigned to device CPU, is_swa = 0 +load_tensors: layer 30 assigned to device CPU, is_swa = 0 +load_tensors: layer 31 assigned to device CPU, is_swa = 0 +load_tensors: layer 32 assigned to device CPU, is_swa = 0 +load_tensors: layer 33 assigned to device CPU, is_swa = 0 +load_tensors: layer 34 assigned to device CPU, is_swa = 0 +load_tensors: layer 35 assigned to device CPU, is_swa = 0 +load_tensors: layer 36 assigned to device CPU, is_swa = 0 +load_tensors: layer 37 assigned to device CPU, is_swa = 0 +load_tensors: layer 38 assigned to device CPU, is_swa = 0 +load_tensors: layer 39 assigned to device CPU, is_swa = 0 +load_tensors: layer 40 assigned to device CPU, is_swa = 0 +load_tensors: layer 41 assigned to device CPU, is_swa = 0 +load_tensors: layer 42 assigned to device CPU, is_swa = 0 +load_tensors: layer 43 assigned to device CPU, is_swa = 0 +load_tensors: layer 44 assigned to device CPU, is_swa = 0 +load_tensors: layer 45 assigned to device CPU, is_swa = 0 +load_tensors: layer 46 assigned to device CPU, is_swa = 0 +load_tensors: layer 47 assigned to device CPU, is_swa = 0 +load_tensors: layer 48 assigned to device CPU, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +create_tensor: loading tensor blk.23.ffn_down_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +create_tensor: loading tensor blk.24.ffn_down_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +create_tensor: loading tensor blk.25.ffn_down_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +create_tensor: loading tensor blk.26.ffn_down_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +create_tensor: loading tensor blk.27.ffn_down_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +create_tensor: loading tensor blk.28.ffn_down_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +create_tensor: loading tensor blk.29.ffn_down_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +create_tensor: loading tensor blk.30.ffn_down_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +create_tensor: loading tensor blk.31.ffn_down_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +create_tensor: loading tensor blk.32.ffn_down_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +create_tensor: loading tensor blk.33.ffn_down_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +create_tensor: loading tensor blk.34.ffn_down_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +create_tensor: loading tensor blk.35.ffn_down_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +create_tensor: loading tensor blk.36.ffn_down_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +create_tensor: loading tensor blk.37.ffn_down_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +create_tensor: loading tensor blk.38.ffn_down_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +create_tensor: loading tensor blk.39.ffn_down_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +create_tensor: loading tensor blk.40.ffn_down_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +create_tensor: loading tensor blk.41.ffn_down_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +create_tensor: loading tensor blk.42.ffn_down_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +create_tensor: loading tensor blk.43.ffn_down_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +create_tensor: loading tensor blk.44.ffn_down_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +create_tensor: loading tensor blk.45.ffn_down_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +create_tensor: loading tensor blk.46.ffn_down_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +create_tensor: loading tensor blk.47.ffn_down_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 0 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading 0 repeating layers to GPU +load_tensors: offloaded 0/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm_Host model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: CPU output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = CPU +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = CPU +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = CPU +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = CPU +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = CPU +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = CPU +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = CPU +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = CPU +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = CPU +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = CPU +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = CPU +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = CPU +llama_kv_cache: CPU KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = CPU +llama_memory_recurrent, layer 1: dev = CPU +llama_memory_recurrent, layer 2: dev = CPU +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = CPU +llama_memory_recurrent, layer 5: dev = CPU +llama_memory_recurrent, layer 6: dev = CPU +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = CPU +llama_memory_recurrent, layer 9: dev = CPU +llama_memory_recurrent, layer 10: dev = CPU +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = CPU +llama_memory_recurrent, layer 13: dev = CPU +llama_memory_recurrent, layer 14: dev = CPU +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = CPU +llama_memory_recurrent, layer 17: dev = CPU +llama_memory_recurrent, layer 18: dev = CPU +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = CPU +llama_memory_recurrent, layer 21: dev = CPU +llama_memory_recurrent, layer 22: dev = CPU +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = CPU +llama_memory_recurrent, layer 25: dev = CPU +llama_memory_recurrent, layer 26: dev = CPU +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = CPU +llama_memory_recurrent, layer 29: dev = CPU +llama_memory_recurrent, layer 30: dev = CPU +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = CPU +llama_memory_recurrent, layer 33: dev = CPU +llama_memory_recurrent, layer 34: dev = CPU +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = CPU +llama_memory_recurrent, layer 37: dev = CPU +llama_memory_recurrent, layer 38: dev = CPU +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = CPU +llama_memory_recurrent, layer 41: dev = CPU +llama_memory_recurrent, layer 42: dev = CPU +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = CPU +llama_memory_recurrent, layer 45: dev = CPU +llama_memory_recurrent, layer 46: dev = CPU +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: CPU RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 1099.00 MiB +sched_reserve: ROCm_Host compute buffer size = 276.11 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 976 (with bs=512), 73 (with bs=1) +sched_reserve: reserve took 6.97 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32586 + ( 1099 = 0 + 0 + 1099) + 17592186043483 | +llama_memory_breakdown_print: | - Host | 57632 = 54208 + 3147 + 276 | +llama_params_fit_impl: memory for test allocation by device: +llama_params_fit_impl: id=0, n_layer= 0, n_part= 0, overflow_type=4, mem= 1099 MiB +llama_params_fit_impl: filling dense-only layers back-to-front: +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +tensor blk.1.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.1.ffn_down_exps.weight +tensor blk.1.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +tensor blk.1.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +tensor blk.2.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.2.ffn_down_exps.weight +tensor blk.2.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +tensor blk.2.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +tensor blk.3.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.3.ffn_down_exps.weight +tensor blk.3.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +tensor blk.3.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +tensor blk.4.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.4.ffn_down_exps.weight +tensor blk.4.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +tensor blk.4.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +tensor blk.5.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.5.ffn_down_exps.weight +tensor blk.5.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +tensor blk.5.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +tensor blk.6.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.6.ffn_down_exps.weight +tensor blk.6.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +tensor blk.6.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +tensor blk.7.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.7.ffn_down_exps.weight +tensor blk.7.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +tensor blk.7.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +tensor blk.8.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.8.ffn_down_exps.weight +tensor blk.8.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +tensor blk.8.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +tensor blk.9.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.9.ffn_down_exps.weight +tensor blk.9.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +tensor blk.9.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +tensor blk.10.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.10.ffn_down_exps.weight +tensor blk.10.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +tensor blk.10.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +tensor blk.11.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.11.ffn_down_exps.weight +tensor blk.11.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +tensor blk.11.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +tensor blk.12.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.12.ffn_down_exps.weight +tensor blk.12.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +tensor blk.12.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +tensor blk.13.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.13.ffn_down_exps.weight +tensor blk.13.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +tensor blk.13.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +tensor blk.14.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.14.ffn_down_exps.weight +tensor blk.14.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +tensor blk.14.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +tensor blk.15.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.15.ffn_down_exps.weight +tensor blk.15.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +tensor blk.15.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +tensor blk.16.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.16.ffn_down_exps.weight +tensor blk.16.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +tensor blk.16.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +tensor blk.17.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.17.ffn_down_exps.weight +tensor blk.17.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +tensor blk.17.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +tensor blk.18.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.18.ffn_down_exps.weight +tensor blk.18.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +tensor blk.18.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +tensor blk.19.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.19.ffn_down_exps.weight +tensor blk.19.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +tensor blk.19.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +tensor blk.20.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.20.ffn_down_exps.weight +tensor blk.20.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +tensor blk.20.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +tensor blk.21.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.21.ffn_down_exps.weight +tensor blk.21.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +tensor blk.21.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +tensor blk.22.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.22.ffn_down_exps.weight +tensor blk.22.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +tensor blk.22.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +tensor blk.23.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_exps.weight +tensor blk.23.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +tensor blk.23.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +tensor blk.24.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_down_exps.weight +tensor blk.24.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +tensor blk.24.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +tensor blk.25.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_down_exps.weight +tensor blk.25.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +tensor blk.25.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +tensor blk.26.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_down_exps.weight +tensor blk.26.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +tensor blk.26.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +tensor blk.27.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_down_exps.weight +tensor blk.27.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +tensor blk.27.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +tensor blk.28.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_down_exps.weight +tensor blk.28.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +tensor blk.28.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +tensor blk.29.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_down_exps.weight +tensor blk.29.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +tensor blk.29.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +tensor blk.30.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_down_exps.weight +tensor blk.30.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +tensor blk.30.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +tensor blk.31.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_down_exps.weight +tensor blk.31.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +tensor blk.31.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +tensor blk.32.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_down_exps.weight +tensor blk.32.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +tensor blk.32.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +tensor blk.33.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_down_exps.weight +tensor blk.33.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +tensor blk.33.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +tensor blk.34.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_down_exps.weight +tensor blk.34.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +tensor blk.34.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +tensor blk.35.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_down_exps.weight +tensor blk.35.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +tensor blk.35.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +tensor blk.36.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_down_exps.weight +tensor blk.36.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +tensor blk.36.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +tensor blk.37.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_down_exps.weight +tensor blk.37.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +tensor blk.37.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +tensor blk.38.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_down_exps.weight +tensor blk.38.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +tensor blk.38.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +tensor blk.39.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_down_exps.weight +tensor blk.39.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +tensor blk.39.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +tensor blk.40.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_down_exps.weight +tensor blk.40.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +tensor blk.40.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +tensor blk.41.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_down_exps.weight +tensor blk.41.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +tensor blk.41.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +tensor blk.42.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_down_exps.weight +tensor blk.42.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +tensor blk.42.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +tensor blk.43.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_down_exps.weight +tensor blk.43.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +tensor blk.43.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +tensor blk.44.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_down_exps.weight +tensor blk.44.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +tensor blk.44.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +tensor blk.45.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_down_exps.weight +tensor blk.45.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +tensor blk.45.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +tensor blk.46.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_down_exps.weight +tensor blk.46.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +tensor blk.46.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +tensor blk.47.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_down_exps.weight +tensor blk.47.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +tensor blk.47.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 141 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +load_tensors: ROCm_Host model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 736.00 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 143 (with bs=512), 96 (with bs=1) +sched_reserve: reserve took 5.99 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + ( 6692 = 2808 + 3147 + 736) + 17592186037965 | +llama_memory_breakdown_print: | - Host | 51664 = 51400 + 0 + 264 | +llama_params_fit_impl: memory for test allocation by device: +llama_params_fit_impl: id=0, n_layer=49, n_part=48, overflow_type=4, mem= 6692 MiB +llama_params_fit_impl: set ngl_per_device[0].n_layer=49 +llama_params_fit_impl: - ROCm0 (AMD Instinct MI100): 49 layers, 6692 MiB used, 25817 MiB free +llama_params_fit_impl: converting dense-only layers to full layers and filling them front-to-back with overflow to next device/system memory: +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +create_tensor: loading tensor blk.23.ffn_down_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +create_tensor: loading tensor blk.24.ffn_down_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +create_tensor: loading tensor blk.25.ffn_down_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +create_tensor: loading tensor blk.26.ffn_down_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +create_tensor: loading tensor blk.27.ffn_down_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +create_tensor: loading tensor blk.28.ffn_down_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +create_tensor: loading tensor blk.29.ffn_down_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +create_tensor: loading tensor blk.30.ffn_down_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +create_tensor: loading tensor blk.31.ffn_down_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +create_tensor: loading tensor blk.32.ffn_down_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +create_tensor: loading tensor blk.33.ffn_down_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +create_tensor: loading tensor blk.34.ffn_down_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +create_tensor: loading tensor blk.35.ffn_down_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +create_tensor: loading tensor blk.36.ffn_down_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +create_tensor: loading tensor blk.37.ffn_down_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +create_tensor: loading tensor blk.38.ffn_down_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +create_tensor: loading tensor blk.39.ffn_down_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +create_tensor: loading tensor blk.40.ffn_down_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +create_tensor: loading tensor blk.41.ffn_down_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +create_tensor: loading tensor blk.42.ffn_down_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +create_tensor: loading tensor blk.43.ffn_down_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +create_tensor: loading tensor blk.44.ffn_down_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +create_tensor: loading tensor blk.45.ffn_down_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +create_tensor: loading tensor blk.46.ffn_down_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +create_tensor: loading tensor blk.47.ffn_down_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 0 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 420.01 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 2 +sched_reserve: reserve took 5.72 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + (57572 = 54004 + 3147 + 420) + 17592185987085 | +llama_memory_breakdown_print: | - Host | 468 = 204 + 0 + 264 | +llama_params_fit_impl: memory for test allocation by device: +llama_params_fit_impl: id=0, n_layer=49, n_part= 0, overflow_type=4, mem= 57572 MiB +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +create_tensor: loading tensor blk.23.ffn_down_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +tensor blk.24.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_down_exps.weight +tensor blk.24.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +tensor blk.24.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +tensor blk.25.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_down_exps.weight +tensor blk.25.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +tensor blk.25.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +tensor blk.26.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_down_exps.weight +tensor blk.26.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +tensor blk.26.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +tensor blk.27.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_down_exps.weight +tensor blk.27.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +tensor blk.27.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +tensor blk.28.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_down_exps.weight +tensor blk.28.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +tensor blk.28.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +tensor blk.29.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_down_exps.weight +tensor blk.29.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +tensor blk.29.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +tensor blk.30.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_down_exps.weight +tensor blk.30.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +tensor blk.30.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +tensor blk.31.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_down_exps.weight +tensor blk.31.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +tensor blk.31.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +tensor blk.32.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_down_exps.weight +tensor blk.32.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +tensor blk.32.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +tensor blk.33.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_down_exps.weight +tensor blk.33.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +tensor blk.33.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +tensor blk.34.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_down_exps.weight +tensor blk.34.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +tensor blk.34.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +tensor blk.35.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_down_exps.weight +tensor blk.35.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +tensor blk.35.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +tensor blk.36.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_down_exps.weight +tensor blk.36.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +tensor blk.36.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +tensor blk.37.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_down_exps.weight +tensor blk.37.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +tensor blk.37.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +tensor blk.38.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_down_exps.weight +tensor blk.38.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +tensor blk.38.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +tensor blk.39.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_down_exps.weight +tensor blk.39.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +tensor blk.39.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +tensor blk.40.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_down_exps.weight +tensor blk.40.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +tensor blk.40.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +tensor blk.41.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_down_exps.weight +tensor blk.41.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +tensor blk.41.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +tensor blk.42.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_down_exps.weight +tensor blk.42.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +tensor blk.42.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +tensor blk.43.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_down_exps.weight +tensor blk.43.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +tensor blk.43.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +tensor blk.44.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_down_exps.weight +tensor blk.44.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +tensor blk.44.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +tensor blk.45.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_down_exps.weight +tensor blk.45.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +tensor blk.45.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +tensor blk.46.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_down_exps.weight +tensor blk.46.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +tensor blk.46.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +tensor blk.47.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_down_exps.weight +tensor blk.47.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +tensor blk.47.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 72 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +load_tensors: ROCm_Host model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 840.01 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 74 (with bs=512), 50 (with bs=1) +sched_reserve: reserve took 5.86 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + (31832 = 27844 + 3147 + 840) + 17592186012825 | +llama_memory_breakdown_print: | - Host | 26628 = 26364 + 0 + 264 | +llama_params_fit_impl: memory for test allocation by device: +llama_params_fit_impl: id=0, n_layer=49, n_part=25, overflow_type=4, mem= 31832 MiB +llama_params_fit_impl: set ngl_per_device_high[0].(n_layer, n_part)=(49, 25), id_dense_start_high=0 +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +tensor blk.23.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_exps.weight +tensor blk.23.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +tensor blk.23.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +tensor blk.24.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_down_exps.weight +tensor blk.24.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +tensor blk.24.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +tensor blk.25.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_down_exps.weight +tensor blk.25.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +tensor blk.25.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +tensor blk.26.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_down_exps.weight +tensor blk.26.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +tensor blk.26.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +tensor blk.27.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_down_exps.weight +tensor blk.27.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +tensor blk.27.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +tensor blk.28.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_down_exps.weight +tensor blk.28.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +tensor blk.28.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +tensor blk.29.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_down_exps.weight +tensor blk.29.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +tensor blk.29.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +tensor blk.30.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_down_exps.weight +tensor blk.30.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +tensor blk.30.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +tensor blk.31.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_down_exps.weight +tensor blk.31.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +tensor blk.31.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +tensor blk.32.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_down_exps.weight +tensor blk.32.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +tensor blk.32.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +tensor blk.33.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_down_exps.weight +tensor blk.33.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +tensor blk.33.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +tensor blk.34.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_down_exps.weight +tensor blk.34.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +tensor blk.34.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +tensor blk.35.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_down_exps.weight +tensor blk.35.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +tensor blk.35.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +tensor blk.36.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_down_exps.weight +tensor blk.36.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +tensor blk.36.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +tensor blk.37.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_down_exps.weight +tensor blk.37.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +tensor blk.37.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +tensor blk.38.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_down_exps.weight +tensor blk.38.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +tensor blk.38.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +tensor blk.39.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_down_exps.weight +tensor blk.39.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +tensor blk.39.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +tensor blk.40.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_down_exps.weight +tensor blk.40.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +tensor blk.40.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +tensor blk.41.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_down_exps.weight +tensor blk.41.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +tensor blk.41.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +tensor blk.42.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_down_exps.weight +tensor blk.42.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +tensor blk.42.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +tensor blk.43.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_down_exps.weight +tensor blk.43.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +tensor blk.43.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +tensor blk.44.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_down_exps.weight +tensor blk.44.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +tensor blk.44.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +tensor blk.45.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_down_exps.weight +tensor blk.45.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +tensor blk.45.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +tensor blk.46.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_down_exps.weight +tensor blk.46.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +tensor blk.46.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +tensor blk.47.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_down_exps.weight +tensor blk.47.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +tensor blk.47.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 75 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +load_tensors: ROCm_Host model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 840.01 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 77 (with bs=512), 52 (with bs=1) +sched_reserve: reserve took 6.19 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + (30708 = 26720 + 3147 + 840) + 17592186013949 | +llama_memory_breakdown_print: | - Host | 27752 = 27488 + 0 + 264 | +llama_params_fit_impl: memory for test allocation by device: +llama_params_fit_impl: id=0, n_layer=49, n_part=26, overflow_type=4, mem= 30708 MiB +llama_params_fit_impl: set ngl_per_device[0].(n_layer, n_part)=(49, 26), id_dense_start=0 +llama_params_fit_impl: trying to fit one extra layer with overflow_type=LAYER_FRACTION_UP +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +tensor blk.23.ffn_gate_inp.weight (4 MiB f32) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +tensor blk.23.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_exps.weight +tensor blk.23.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +create_tensor: loading tensor blk.23.ffn_up_exps.weight +tensor blk.23.ffn_gate_inp_shexp.weight (0 MiB f32) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +tensor blk.23.ffn_gate_shexp.weight (0 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +tensor blk.23.ffn_down_shexp.weight (0 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +tensor blk.24.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_down_exps.weight +tensor blk.24.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +tensor blk.24.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +tensor blk.25.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_down_exps.weight +tensor blk.25.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +tensor blk.25.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +tensor blk.26.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_down_exps.weight +tensor blk.26.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +tensor blk.26.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +tensor blk.27.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_down_exps.weight +tensor blk.27.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +tensor blk.27.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +tensor blk.28.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_down_exps.weight +tensor blk.28.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +tensor blk.28.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +tensor blk.29.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_down_exps.weight +tensor blk.29.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +tensor blk.29.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +tensor blk.30.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_down_exps.weight +tensor blk.30.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +tensor blk.30.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +tensor blk.31.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_down_exps.weight +tensor blk.31.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +tensor blk.31.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +tensor blk.32.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_down_exps.weight +tensor blk.32.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +tensor blk.32.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +tensor blk.33.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_down_exps.weight +tensor blk.33.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +tensor blk.33.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +tensor blk.34.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_down_exps.weight +tensor blk.34.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +tensor blk.34.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +tensor blk.35.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_down_exps.weight +tensor blk.35.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +tensor blk.35.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +tensor blk.36.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_down_exps.weight +tensor blk.36.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +tensor blk.36.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +tensor blk.37.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_down_exps.weight +tensor blk.37.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +tensor blk.37.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +tensor blk.38.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_down_exps.weight +tensor blk.38.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +tensor blk.38.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +tensor blk.39.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_down_exps.weight +tensor blk.39.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +tensor blk.39.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +tensor blk.40.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_down_exps.weight +tensor blk.40.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +tensor blk.40.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +tensor blk.41.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_down_exps.weight +tensor blk.41.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +tensor blk.41.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +tensor blk.42.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_down_exps.weight +tensor blk.42.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +tensor blk.42.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +tensor blk.43.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_down_exps.weight +tensor blk.43.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +tensor blk.43.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +tensor blk.44.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_down_exps.weight +tensor blk.44.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +tensor blk.44.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +tensor blk.45.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_down_exps.weight +tensor blk.45.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +tensor blk.45.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +tensor blk.46.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_down_exps.weight +tensor blk.46.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +tensor blk.46.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +tensor blk.47.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_down_exps.weight +tensor blk.47.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +tensor blk.47.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 78 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +load_tensors: ROCm_Host model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 840.01 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 80 (with bs=512), 56 (with bs=1) +sched_reserve: reserve took 5.88 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + (31054 = 27066 + 3147 + 840) + 17592186013603 | +llama_memory_breakdown_print: | - Host | 27405 = 27141 + 0 + 264 | +llama_params_fit_impl: memory for test allocation by device: +llama_params_fit_impl: id=0, n_layer=49, n_part=26, overflow_type=2, mem= 31054 MiB +llama_params_fit_impl: set ngl_per_device[0].(n_layer, n_part, overflow_type)=(49, 26, UP), id_dense_start=0 +llama_params_fit_impl: trying to fit one extra layer with overflow_type=LAYER_FRACTION_GATE +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 1 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +tensor blk.23.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +tensor blk.23.ffn_down_shexp.weight (0 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +tensor blk.24.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_down_exps.weight +tensor blk.24.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +tensor blk.24.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +tensor blk.25.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_down_exps.weight +tensor blk.25.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +tensor blk.25.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +tensor blk.26.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_down_exps.weight +tensor blk.26.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +tensor blk.26.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +tensor blk.27.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_down_exps.weight +tensor blk.27.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +tensor blk.27.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +tensor blk.28.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_down_exps.weight +tensor blk.28.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +tensor blk.28.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +tensor blk.29.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_down_exps.weight +tensor blk.29.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +tensor blk.29.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +tensor blk.30.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_down_exps.weight +tensor blk.30.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +tensor blk.30.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +tensor blk.31.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_down_exps.weight +tensor blk.31.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +tensor blk.31.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +tensor blk.32.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_down_exps.weight +tensor blk.32.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +tensor blk.32.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +tensor blk.33.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_down_exps.weight +tensor blk.33.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +tensor blk.33.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +tensor blk.34.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_down_exps.weight +tensor blk.34.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +tensor blk.34.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +tensor blk.35.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_down_exps.weight +tensor blk.35.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +tensor blk.35.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +tensor blk.36.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_down_exps.weight +tensor blk.36.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +tensor blk.36.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +tensor blk.37.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_down_exps.weight +tensor blk.37.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +tensor blk.37.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +tensor blk.38.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_down_exps.weight +tensor blk.38.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +tensor blk.38.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +tensor blk.39.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_down_exps.weight +tensor blk.39.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +tensor blk.39.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +tensor blk.40.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_down_exps.weight +tensor blk.40.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +tensor blk.40.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +tensor blk.41.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_down_exps.weight +tensor blk.41.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +tensor blk.41.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +tensor blk.42.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_down_exps.weight +tensor blk.42.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +tensor blk.42.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +tensor blk.43.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_down_exps.weight +tensor blk.43.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +tensor blk.43.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +tensor blk.44.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_down_exps.weight +tensor blk.44.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +tensor blk.44.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +tensor blk.45.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_down_exps.weight +tensor blk.45.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +tensor blk.45.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +tensor blk.46.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_down_exps.weight +tensor blk.46.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +tensor blk.46.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +tensor blk.47.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_down_exps.weight +tensor blk.47.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +tensor blk.47.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 74 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 0.00 MiB +load_tensors: ROCm0 model buffer size = 0.00 MiB +load_tensors: ROCm_Host model buffer size = 0.00 MiB +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 0.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 840.01 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 76 (with bs=512), 54 (with bs=1) +sched_reserve: reserve took 5.75 ms, sched copies = 1 +llama_memory_breakdown_print: | memory breakdown [MiB] | total free self model context compute unaccounted | +llama_memory_breakdown_print: | - ROCm0 (MI100) | 32752 = 32510 + (31411 = 27423 + 3147 + 840) + 17592186013246 | +llama_memory_breakdown_print: | - Host | 27048 = 26784 + 0 + 264 | +llama_params_fit_impl: memory for test allocation by device: +llama_params_fit_impl: id=0, n_layer=49, n_part=26, overflow_type=3, mem= 31411 MiB +llama_params_fit_impl: set ngl_per_device[0].(n_layer, n_part, overflow_type)=(49, 26, GATE), id_dense_start=0 +llama_params_fit_impl: - ROCm0 (AMD Instinct MI100): 49 layers (26 overflowing), 31411 MiB used, 1098 MiB free +llama_params_fit: successfully fit params to free device memory +llama_params_fit: fitting params to free memory took 1.44 seconds +llama_model_load_from_file_impl: using device ROCm0 (AMD Instinct MI100) (0000:03:00.0) - 32586 MiB free +llama_model_loader: additional 2 GGUFs metadata loaded. +llama_model_loader: loaded meta data with 56 key-value pairs and 843 tensors from /home/edwlan/.cache/llama.cpp/unsloth_Qwen3-Coder-Next-GGUF_Q5_K_M_Qwen3-Coder-Next-Q5_K_M-00001-of-00003.gguf (version GGUF V3 (latest)) +llama_model_loader: Dumping metadata keys/values. Note: KV overrides do not apply in this output. +llama_model_loader: - kv 0: general.architecture str = qwen3next +llama_model_loader: - kv 1: general.type str = model +llama_model_loader: - kv 2: general.sampling.top_k i32 = 40 +llama_model_loader: - kv 3: general.sampling.top_p f32 = 0.950000 +llama_model_loader: - kv 4: general.sampling.temp f32 = 1.000000 +llama_model_loader: - kv 5: general.name str = Qwen3-Coder-Next +llama_model_loader: - kv 6: general.basename str = Qwen3-Coder-Next +llama_model_loader: - kv 7: general.quantized_by str = Unsloth +llama_model_loader: - kv 8: general.size_label str = 512x2.5B +llama_model_loader: - kv 9: general.license str = apache-2.0 +llama_model_loader: - kv 10: general.license.link str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 11: general.repo_url str = https://huggingface.co/unsloth +llama_model_loader: - kv 12: general.base_model.count u32 = 1 +llama_model_loader: - kv 13: general.base_model.0.name str = Qwen3 Coder Next +llama_model_loader: - kv 14: general.base_model.0.organization str = Qwen +llama_model_loader: - kv 15: general.base_model.0.repo_url str = https://huggingface.co/Qwen/Qwen3-Cod... +llama_model_loader: - kv 16: general.tags arr[str,2] = ["unsloth", "text-generation"] +llama_model_loader: - kv 17: qwen3next.block_count u32 = 48 +llama_model_loader: - kv 18: qwen3next.context_length u32 = 262144 +llama_model_loader: - kv 19: qwen3next.embedding_length u32 = 2048 +llama_model_loader: - kv 20: qwen3next.feed_forward_length u32 = 5120 +llama_model_loader: - kv 21: qwen3next.attention.head_count u32 = 16 +llama_model_loader: - kv 22: qwen3next.attention.head_count_kv u32 = 2 +llama_model_loader: - kv 23: qwen3next.rope.freq_base f32 = 5000000.000000 +llama_model_loader: - kv 24: qwen3next.attention.layer_norm_rms_epsilon f32 = 0.000001 +llama_model_loader: - kv 25: qwen3next.expert_count u32 = 512 +llama_model_loader: - kv 26: qwen3next.expert_used_count u32 = 10 +llama_model_loader: - kv 27: qwen3next.attention.key_length u32 = 256 +llama_model_loader: - kv 28: qwen3next.attention.value_length u32 = 256 +llama_model_loader: - kv 29: qwen3next.expert_feed_forward_length u32 = 512 +llama_model_loader: - kv 30: qwen3next.expert_shared_feed_forward_length u32 = 512 +llama_model_loader: - kv 31: qwen3next.ssm.conv_kernel u32 = 4 +llama_model_loader: - kv 32: qwen3next.ssm.state_size u32 = 128 +llama_model_loader: - kv 33: qwen3next.ssm.group_count u32 = 16 +llama_model_loader: - kv 34: qwen3next.ssm.time_step_rank u32 = 32 +llama_model_loader: - kv 35: qwen3next.ssm.inner_size u32 = 4096 +llama_model_loader: - kv 36: qwen3next.full_attention_interval u32 = 4 +llama_model_loader: - kv 37: qwen3next.rope.dimension_count u32 = 64 +llama_model_loader: - kv 38: tokenizer.ggml.model str = gpt2 +llama_model_loader: - kv 39: tokenizer.ggml.pre str = qwen2 +llama_model_loader: - kv 40: tokenizer.ggml.tokens arr[str,151936] = ["!", "\"", "#", "$", "%", "&", "'", ... +llama_model_loader: - kv 41: tokenizer.ggml.token_type arr[i32,151936] = [1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, ... +llama_model_loader: - kv 42: tokenizer.ggml.merges arr[str,151387] = ["Ġ Ġ", "ĠĠ ĠĠ", "i n", "Ġ t",... +llama_model_loader: - kv 43: tokenizer.ggml.eos_token_id u32 = 151645 +llama_model_loader: - kv 44: tokenizer.ggml.padding_token_id u32 = 151654 +llama_model_loader: - kv 45: tokenizer.ggml.add_bos_token bool = false +llama_model_loader: - kv 46: tokenizer.chat_template str = {% macro render_extra_keys(json_dict,... +llama_model_loader: - kv 47: general.quantization_version u32 = 2 +llama_model_loader: - kv 48: general.file_type u32 = 17 +llama_model_loader: - kv 49: quantize.imatrix.file str = Qwen3-Coder-Next-GGUF/imatrix_unsloth... +llama_model_loader: - kv 50: quantize.imatrix.dataset str = unsloth_calibration_Qwen3-Coder-Next.txt +llama_model_loader: - kv 51: quantize.imatrix.entries_count u32 = 576 +llama_model_loader: - kv 52: quantize.imatrix.chunks_count u32 = 154 +llama_model_loader: - kv 53: split.no u16 = 0 +llama_model_loader: - kv 54: split.tensors.count i32 = 843 +llama_model_loader: - kv 55: split.count u16 = 3 +llama_model_loader: - type f32: 361 tensors +llama_model_loader: - type q5_K: 233 tensors +llama_model_loader: - type q6_K: 249 tensors +print_info: file format = GGUF V3 (latest) +print_info: file type = Q5_K - Medium +print_info: file size = 52.94 GiB (5.71 BPW) +init_tokenizer: initializing tokenizer for type 2 +load: 0 unused tokens +load: control token: 151660 '<|fim_middle|>' is not marked as EOG +load: control token: 151659 '<|fim_prefix|>' is not marked as EOG +load: control token: 151653 '<|vision_end|>' is not marked as EOG +load: control token: 151648 '<|box_start|>' is not marked as EOG +load: control token: 151646 '<|object_ref_start|>' is not marked as EOG +load: control token: 151649 '<|box_end|>' is not marked as EOG +load: control-looking token: 128247 '' was not control-type; this is probably a bug in the model. its type will be overridden +load: control token: 151655 '<|image_pad|>' is not marked as EOG +load: control token: 151651 '<|quad_end|>' is not marked as EOG +load: control token: 151647 '<|object_ref_end|>' is not marked as EOG +load: control token: 151652 '<|vision_start|>' is not marked as EOG +load: control token: 151654 '<|vision_pad|>' is not marked as EOG +load: control token: 151656 '<|video_pad|>' is not marked as EOG +load: control token: 151644 '<|im_start|>' is not marked as EOG +load: control token: 151661 '<|fim_suffix|>' is not marked as EOG +load: control token: 151650 '<|quad_start|>' is not marked as EOG +load: printing all EOG tokens: +load: - 128247 ('') +load: - 151643 ('<|endoftext|>') +load: - 151645 ('<|im_end|>') +load: - 151662 ('<|fim_pad|>') +load: - 151663 ('<|repo_name|>') +load: - 151664 ('<|file_sep|>') +load: special tokens cache size = 27 +load: token to piece cache size = 0.9311 MB +print_info: arch = qwen3next +print_info: vocab_only = 0 +print_info: no_alloc = 0 +print_info: n_ctx_train = 262144 +print_info: n_embd = 2048 +print_info: n_embd_inp = 2048 +print_info: n_layer = 48 +print_info: n_head = 16 +print_info: n_head_kv = 2 +print_info: n_rot = 64 +print_info: n_swa = 0 +print_info: is_swa_any = 0 +print_info: n_embd_head_k = 256 +print_info: n_embd_head_v = 256 +print_info: n_gqa = 8 +print_info: n_embd_k_gqa = 512 +print_info: n_embd_v_gqa = 512 +print_info: f_norm_eps = 0.0e+00 +print_info: f_norm_rms_eps = 1.0e-06 +print_info: f_clamp_kqv = 0.0e+00 +print_info: f_max_alibi_bias = 0.0e+00 +print_info: f_logit_scale = 0.0e+00 +print_info: f_attn_scale = 0.0e+00 +print_info: n_ff = 5120 +print_info: n_expert = 512 +print_info: n_expert_used = 10 +print_info: n_expert_groups = 0 +print_info: n_group_used = 0 +print_info: causal attn = 1 +print_info: pooling type = 0 +print_info: rope type = 2 +print_info: rope scaling = linear +print_info: freq_base_train = 5000000.0 +print_info: freq_scale_train = 1 +print_info: n_ctx_orig_yarn = 262144 +print_info: rope_yarn_log_mul = 0.0000 +print_info: rope_finetuned = unknown +print_info: ssm_d_conv = 4 +print_info: ssm_d_inner = 4096 +print_info: ssm_d_state = 128 +print_info: ssm_dt_rank = 32 +print_info: ssm_n_group = 16 +print_info: ssm_dt_b_c_rms = 0 +print_info: model type = 80B.A3B +print_info: model params = 79.67 B +print_info: general.name = Qwen3-Coder-Next +print_info: vocab type = BPE +print_info: n_vocab = 151936 +print_info: n_merges = 151387 +print_info: BOS token = 11 ',' +print_info: EOS token = 151645 '<|im_end|>' +print_info: EOT token = 151645 '<|im_end|>' +print_info: PAD token = 151654 '<|vision_pad|>' +print_info: LF token = 198 'Ċ' +print_info: FIM PRE token = 151659 '<|fim_prefix|>' +print_info: FIM SUF token = 151661 '<|fim_suffix|>' +print_info: FIM MID token = 151660 '<|fim_middle|>' +print_info: FIM PAD token = 151662 '<|fim_pad|>' +print_info: FIM REP token = 151663 '<|repo_name|>' +print_info: FIM SEP token = 151664 '<|file_sep|>' +print_info: EOG token = 128247 '' +print_info: EOG token = 151643 '<|endoftext|>' +print_info: EOG token = 151645 '<|im_end|>' +print_info: EOG token = 151662 '<|fim_pad|>' +print_info: EOG token = 151663 '<|repo_name|>' +print_info: EOG token = 151664 '<|file_sep|>' +print_info: max token length = 256 +load_tensors: loading model tensors, this can take a while... (mmap = false, direct_io = false) +load_tensors: layer 0 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 1 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 2 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 3 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 4 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 5 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 6 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 7 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 8 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 9 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 10 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 11 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 12 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 13 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 14 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 15 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 16 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 17 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 18 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 19 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 20 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 21 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 22 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 23 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 24 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 25 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 26 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 27 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 28 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 29 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 30 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 31 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 32 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 33 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 34 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 35 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 36 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 37 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 38 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 39 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 40 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 41 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 42 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 43 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 44 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 45 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 46 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 47 assigned to device ROCm0, is_swa = 0 +load_tensors: layer 48 assigned to device ROCm0, is_swa = 0 +create_tensor: loading tensor token_embd.weight +create_tensor: loading tensor output_norm.weight +create_tensor: loading tensor output.weight +create_tensor: loading tensor blk.0.attn_norm.weight +create_tensor: loading tensor blk.0.post_attention_norm.weight +create_tensor: loading tensor blk.0.attn_qkv.weight +create_tensor: loading tensor blk.0.attn_gate.weight +create_tensor: loading tensor blk.0.ssm_conv1d.weight +create_tensor: loading tensor blk.0.ssm_dt.bias +create_tensor: loading tensor blk.0.ssm_a +create_tensor: loading tensor blk.0.ssm_ba.weight +create_tensor: loading tensor blk.0.ssm_norm.weight +create_tensor: loading tensor blk.0.ssm_out.weight +create_tensor: loading tensor blk.0.ffn_gate_inp.weight +create_tensor: loading tensor blk.0.ffn_down_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_exps.weight +create_tensor: loading tensor blk.0.ffn_up_exps.weight +create_tensor: loading tensor blk.0.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.0.ffn_gate_shexp.weight +create_tensor: loading tensor blk.0.ffn_up_shexp.weight +create_tensor: loading tensor blk.0.ffn_down_shexp.weight +create_tensor: loading tensor blk.1.attn_norm.weight +create_tensor: loading tensor blk.1.post_attention_norm.weight +create_tensor: loading tensor blk.1.attn_qkv.weight +create_tensor: loading tensor blk.1.attn_gate.weight +create_tensor: loading tensor blk.1.ssm_conv1d.weight +create_tensor: loading tensor blk.1.ssm_dt.bias +create_tensor: loading tensor blk.1.ssm_a +create_tensor: loading tensor blk.1.ssm_ba.weight +create_tensor: loading tensor blk.1.ssm_norm.weight +create_tensor: loading tensor blk.1.ssm_out.weight +create_tensor: loading tensor blk.1.ffn_gate_inp.weight +create_tensor: loading tensor blk.1.ffn_down_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_exps.weight +create_tensor: loading tensor blk.1.ffn_up_exps.weight +create_tensor: loading tensor blk.1.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.1.ffn_gate_shexp.weight +create_tensor: loading tensor blk.1.ffn_up_shexp.weight +create_tensor: loading tensor blk.1.ffn_down_shexp.weight +create_tensor: loading tensor blk.2.attn_norm.weight +create_tensor: loading tensor blk.2.post_attention_norm.weight +create_tensor: loading tensor blk.2.attn_qkv.weight +create_tensor: loading tensor blk.2.attn_gate.weight +create_tensor: loading tensor blk.2.ssm_conv1d.weight +create_tensor: loading tensor blk.2.ssm_dt.bias +create_tensor: loading tensor blk.2.ssm_a +create_tensor: loading tensor blk.2.ssm_ba.weight +create_tensor: loading tensor blk.2.ssm_norm.weight +create_tensor: loading tensor blk.2.ssm_out.weight +create_tensor: loading tensor blk.2.ffn_gate_inp.weight +create_tensor: loading tensor blk.2.ffn_down_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_exps.weight +create_tensor: loading tensor blk.2.ffn_up_exps.weight +create_tensor: loading tensor blk.2.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.2.ffn_gate_shexp.weight +create_tensor: loading tensor blk.2.ffn_up_shexp.weight +create_tensor: loading tensor blk.2.ffn_down_shexp.weight +create_tensor: loading tensor blk.3.attn_norm.weight +create_tensor: loading tensor blk.3.post_attention_norm.weight +create_tensor: loading tensor blk.3.attn_q.weight +create_tensor: loading tensor blk.3.attn_k.weight +create_tensor: loading tensor blk.3.attn_v.weight +create_tensor: loading tensor blk.3.attn_output.weight +create_tensor: loading tensor blk.3.attn_q_norm.weight +create_tensor: loading tensor blk.3.attn_k_norm.weight +create_tensor: loading tensor blk.3.ffn_gate_inp.weight +create_tensor: loading tensor blk.3.ffn_down_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_exps.weight +create_tensor: loading tensor blk.3.ffn_up_exps.weight +create_tensor: loading tensor blk.3.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.3.ffn_gate_shexp.weight +create_tensor: loading tensor blk.3.ffn_up_shexp.weight +create_tensor: loading tensor blk.3.ffn_down_shexp.weight +create_tensor: loading tensor blk.4.attn_norm.weight +create_tensor: loading tensor blk.4.post_attention_norm.weight +create_tensor: loading tensor blk.4.attn_qkv.weight +create_tensor: loading tensor blk.4.attn_gate.weight +create_tensor: loading tensor blk.4.ssm_conv1d.weight +create_tensor: loading tensor blk.4.ssm_dt.bias +create_tensor: loading tensor blk.4.ssm_a +create_tensor: loading tensor blk.4.ssm_ba.weight +create_tensor: loading tensor blk.4.ssm_norm.weight +create_tensor: loading tensor blk.4.ssm_out.weight +create_tensor: loading tensor blk.4.ffn_gate_inp.weight +create_tensor: loading tensor blk.4.ffn_down_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_exps.weight +create_tensor: loading tensor blk.4.ffn_up_exps.weight +create_tensor: loading tensor blk.4.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.4.ffn_gate_shexp.weight +create_tensor: loading tensor blk.4.ffn_up_shexp.weight +create_tensor: loading tensor blk.4.ffn_down_shexp.weight +create_tensor: loading tensor blk.5.attn_norm.weight +create_tensor: loading tensor blk.5.post_attention_norm.weight +create_tensor: loading tensor blk.5.attn_qkv.weight +create_tensor: loading tensor blk.5.attn_gate.weight +create_tensor: loading tensor blk.5.ssm_conv1d.weight +create_tensor: loading tensor blk.5.ssm_dt.bias +create_tensor: loading tensor blk.5.ssm_a +create_tensor: loading tensor blk.5.ssm_ba.weight +create_tensor: loading tensor blk.5.ssm_norm.weight +create_tensor: loading tensor blk.5.ssm_out.weight +create_tensor: loading tensor blk.5.ffn_gate_inp.weight +create_tensor: loading tensor blk.5.ffn_down_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_exps.weight +create_tensor: loading tensor blk.5.ffn_up_exps.weight +create_tensor: loading tensor blk.5.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.5.ffn_gate_shexp.weight +create_tensor: loading tensor blk.5.ffn_up_shexp.weight +create_tensor: loading tensor blk.5.ffn_down_shexp.weight +create_tensor: loading tensor blk.6.attn_norm.weight +create_tensor: loading tensor blk.6.post_attention_norm.weight +create_tensor: loading tensor blk.6.attn_qkv.weight +create_tensor: loading tensor blk.6.attn_gate.weight +create_tensor: loading tensor blk.6.ssm_conv1d.weight +create_tensor: loading tensor blk.6.ssm_dt.bias +create_tensor: loading tensor blk.6.ssm_a +create_tensor: loading tensor blk.6.ssm_ba.weight +create_tensor: loading tensor blk.6.ssm_norm.weight +create_tensor: loading tensor blk.6.ssm_out.weight +create_tensor: loading tensor blk.6.ffn_gate_inp.weight +create_tensor: loading tensor blk.6.ffn_down_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_exps.weight +create_tensor: loading tensor blk.6.ffn_up_exps.weight +create_tensor: loading tensor blk.6.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.6.ffn_gate_shexp.weight +create_tensor: loading tensor blk.6.ffn_up_shexp.weight +create_tensor: loading tensor blk.6.ffn_down_shexp.weight +create_tensor: loading tensor blk.7.attn_norm.weight +create_tensor: loading tensor blk.7.post_attention_norm.weight +create_tensor: loading tensor blk.7.attn_q.weight +create_tensor: loading tensor blk.7.attn_k.weight +create_tensor: loading tensor blk.7.attn_v.weight +create_tensor: loading tensor blk.7.attn_output.weight +create_tensor: loading tensor blk.7.attn_q_norm.weight +create_tensor: loading tensor blk.7.attn_k_norm.weight +create_tensor: loading tensor blk.7.ffn_gate_inp.weight +create_tensor: loading tensor blk.7.ffn_down_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_exps.weight +create_tensor: loading tensor blk.7.ffn_up_exps.weight +create_tensor: loading tensor blk.7.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.7.ffn_gate_shexp.weight +create_tensor: loading tensor blk.7.ffn_up_shexp.weight +create_tensor: loading tensor blk.7.ffn_down_shexp.weight +create_tensor: loading tensor blk.8.attn_norm.weight +create_tensor: loading tensor blk.8.post_attention_norm.weight +create_tensor: loading tensor blk.8.attn_qkv.weight +create_tensor: loading tensor blk.8.attn_gate.weight +create_tensor: loading tensor blk.8.ssm_conv1d.weight +create_tensor: loading tensor blk.8.ssm_dt.bias +create_tensor: loading tensor blk.8.ssm_a +create_tensor: loading tensor blk.8.ssm_ba.weight +create_tensor: loading tensor blk.8.ssm_norm.weight +create_tensor: loading tensor blk.8.ssm_out.weight +create_tensor: loading tensor blk.8.ffn_gate_inp.weight +create_tensor: loading tensor blk.8.ffn_down_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_exps.weight +create_tensor: loading tensor blk.8.ffn_up_exps.weight +create_tensor: loading tensor blk.8.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.8.ffn_gate_shexp.weight +create_tensor: loading tensor blk.8.ffn_up_shexp.weight +create_tensor: loading tensor blk.8.ffn_down_shexp.weight +create_tensor: loading tensor blk.9.attn_norm.weight +create_tensor: loading tensor blk.9.post_attention_norm.weight +create_tensor: loading tensor blk.9.attn_qkv.weight +create_tensor: loading tensor blk.9.attn_gate.weight +create_tensor: loading tensor blk.9.ssm_conv1d.weight +create_tensor: loading tensor blk.9.ssm_dt.bias +create_tensor: loading tensor blk.9.ssm_a +create_tensor: loading tensor blk.9.ssm_ba.weight +create_tensor: loading tensor blk.9.ssm_norm.weight +create_tensor: loading tensor blk.9.ssm_out.weight +create_tensor: loading tensor blk.9.ffn_gate_inp.weight +create_tensor: loading tensor blk.9.ffn_down_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_exps.weight +create_tensor: loading tensor blk.9.ffn_up_exps.weight +create_tensor: loading tensor blk.9.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.9.ffn_gate_shexp.weight +create_tensor: loading tensor blk.9.ffn_up_shexp.weight +create_tensor: loading tensor blk.9.ffn_down_shexp.weight +create_tensor: loading tensor blk.10.attn_norm.weight +create_tensor: loading tensor blk.10.post_attention_norm.weight +create_tensor: loading tensor blk.10.attn_qkv.weight +create_tensor: loading tensor blk.10.attn_gate.weight +create_tensor: loading tensor blk.10.ssm_conv1d.weight +create_tensor: loading tensor blk.10.ssm_dt.bias +create_tensor: loading tensor blk.10.ssm_a +create_tensor: loading tensor blk.10.ssm_ba.weight +create_tensor: loading tensor blk.10.ssm_norm.weight +create_tensor: loading tensor blk.10.ssm_out.weight +create_tensor: loading tensor blk.10.ffn_gate_inp.weight +create_tensor: loading tensor blk.10.ffn_down_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_exps.weight +create_tensor: loading tensor blk.10.ffn_up_exps.weight +create_tensor: loading tensor blk.10.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.10.ffn_gate_shexp.weight +create_tensor: loading tensor blk.10.ffn_up_shexp.weight +create_tensor: loading tensor blk.10.ffn_down_shexp.weight +create_tensor: loading tensor blk.11.attn_norm.weight +create_tensor: loading tensor blk.11.post_attention_norm.weight +create_tensor: loading tensor blk.11.attn_q.weight +create_tensor: loading tensor blk.11.attn_k.weight +create_tensor: loading tensor blk.11.attn_v.weight +create_tensor: loading tensor blk.11.attn_output.weight +create_tensor: loading tensor blk.11.attn_q_norm.weight +create_tensor: loading tensor blk.11.attn_k_norm.weight +create_tensor: loading tensor blk.11.ffn_gate_inp.weight +create_tensor: loading tensor blk.11.ffn_down_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_exps.weight +create_tensor: loading tensor blk.11.ffn_up_exps.weight +create_tensor: loading tensor blk.11.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.11.ffn_gate_shexp.weight +create_tensor: loading tensor blk.11.ffn_up_shexp.weight +create_tensor: loading tensor blk.11.ffn_down_shexp.weight +create_tensor: loading tensor blk.12.attn_norm.weight +create_tensor: loading tensor blk.12.post_attention_norm.weight +create_tensor: loading tensor blk.12.attn_qkv.weight +create_tensor: loading tensor blk.12.attn_gate.weight +create_tensor: loading tensor blk.12.ssm_conv1d.weight +create_tensor: loading tensor blk.12.ssm_dt.bias +create_tensor: loading tensor blk.12.ssm_a +create_tensor: loading tensor blk.12.ssm_ba.weight +create_tensor: loading tensor blk.12.ssm_norm.weight +create_tensor: loading tensor blk.12.ssm_out.weight +create_tensor: loading tensor blk.12.ffn_gate_inp.weight +create_tensor: loading tensor blk.12.ffn_down_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_exps.weight +create_tensor: loading tensor blk.12.ffn_up_exps.weight +create_tensor: loading tensor blk.12.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.12.ffn_gate_shexp.weight +create_tensor: loading tensor blk.12.ffn_up_shexp.weight +create_tensor: loading tensor blk.12.ffn_down_shexp.weight +create_tensor: loading tensor blk.13.attn_norm.weight +create_tensor: loading tensor blk.13.post_attention_norm.weight +create_tensor: loading tensor blk.13.attn_qkv.weight +create_tensor: loading tensor blk.13.attn_gate.weight +create_tensor: loading tensor blk.13.ssm_conv1d.weight +create_tensor: loading tensor blk.13.ssm_dt.bias +create_tensor: loading tensor blk.13.ssm_a +create_tensor: loading tensor blk.13.ssm_ba.weight +create_tensor: loading tensor blk.13.ssm_norm.weight +create_tensor: loading tensor blk.13.ssm_out.weight +create_tensor: loading tensor blk.13.ffn_gate_inp.weight +create_tensor: loading tensor blk.13.ffn_down_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_exps.weight +create_tensor: loading tensor blk.13.ffn_up_exps.weight +create_tensor: loading tensor blk.13.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.13.ffn_gate_shexp.weight +create_tensor: loading tensor blk.13.ffn_up_shexp.weight +create_tensor: loading tensor blk.13.ffn_down_shexp.weight +create_tensor: loading tensor blk.14.attn_norm.weight +create_tensor: loading tensor blk.14.post_attention_norm.weight +create_tensor: loading tensor blk.14.attn_qkv.weight +create_tensor: loading tensor blk.14.attn_gate.weight +create_tensor: loading tensor blk.14.ssm_conv1d.weight +create_tensor: loading tensor blk.14.ssm_dt.bias +create_tensor: loading tensor blk.14.ssm_a +create_tensor: loading tensor blk.14.ssm_ba.weight +create_tensor: loading tensor blk.14.ssm_norm.weight +create_tensor: loading tensor blk.14.ssm_out.weight +create_tensor: loading tensor blk.14.ffn_gate_inp.weight +create_tensor: loading tensor blk.14.ffn_down_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_exps.weight +create_tensor: loading tensor blk.14.ffn_up_exps.weight +create_tensor: loading tensor blk.14.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.14.ffn_gate_shexp.weight +create_tensor: loading tensor blk.14.ffn_up_shexp.weight +create_tensor: loading tensor blk.14.ffn_down_shexp.weight +create_tensor: loading tensor blk.15.attn_norm.weight +create_tensor: loading tensor blk.15.post_attention_norm.weight +create_tensor: loading tensor blk.15.attn_q.weight +create_tensor: loading tensor blk.15.attn_k.weight +create_tensor: loading tensor blk.15.attn_v.weight +create_tensor: loading tensor blk.15.attn_output.weight +create_tensor: loading tensor blk.15.attn_q_norm.weight +create_tensor: loading tensor blk.15.attn_k_norm.weight +create_tensor: loading tensor blk.15.ffn_gate_inp.weight +create_tensor: loading tensor blk.15.ffn_down_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_exps.weight +create_tensor: loading tensor blk.15.ffn_up_exps.weight +create_tensor: loading tensor blk.15.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.15.ffn_gate_shexp.weight +create_tensor: loading tensor blk.15.ffn_up_shexp.weight +create_tensor: loading tensor blk.15.ffn_down_shexp.weight +create_tensor: loading tensor blk.16.attn_norm.weight +create_tensor: loading tensor blk.16.post_attention_norm.weight +create_tensor: loading tensor blk.16.attn_qkv.weight +create_tensor: loading tensor blk.16.attn_gate.weight +create_tensor: loading tensor blk.16.ssm_conv1d.weight +create_tensor: loading tensor blk.16.ssm_dt.bias +create_tensor: loading tensor blk.16.ssm_a +create_tensor: loading tensor blk.16.ssm_ba.weight +create_tensor: loading tensor blk.16.ssm_norm.weight +create_tensor: loading tensor blk.16.ssm_out.weight +create_tensor: loading tensor blk.16.ffn_gate_inp.weight +create_tensor: loading tensor blk.16.ffn_down_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_exps.weight +create_tensor: loading tensor blk.16.ffn_up_exps.weight +create_tensor: loading tensor blk.16.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.16.ffn_gate_shexp.weight +create_tensor: loading tensor blk.16.ffn_up_shexp.weight +create_tensor: loading tensor blk.16.ffn_down_shexp.weight +create_tensor: loading tensor blk.17.attn_norm.weight +create_tensor: loading tensor blk.17.post_attention_norm.weight +create_tensor: loading tensor blk.17.attn_qkv.weight +create_tensor: loading tensor blk.17.attn_gate.weight +create_tensor: loading tensor blk.17.ssm_conv1d.weight +create_tensor: loading tensor blk.17.ssm_dt.bias +create_tensor: loading tensor blk.17.ssm_a +create_tensor: loading tensor blk.17.ssm_ba.weight +create_tensor: loading tensor blk.17.ssm_norm.weight +create_tensor: loading tensor blk.17.ssm_out.weight +create_tensor: loading tensor blk.17.ffn_gate_inp.weight +create_tensor: loading tensor blk.17.ffn_down_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_exps.weight +create_tensor: loading tensor blk.17.ffn_up_exps.weight +create_tensor: loading tensor blk.17.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.17.ffn_gate_shexp.weight +create_tensor: loading tensor blk.17.ffn_up_shexp.weight +create_tensor: loading tensor blk.17.ffn_down_shexp.weight +create_tensor: loading tensor blk.18.attn_norm.weight +create_tensor: loading tensor blk.18.post_attention_norm.weight +create_tensor: loading tensor blk.18.attn_qkv.weight +create_tensor: loading tensor blk.18.attn_gate.weight +create_tensor: loading tensor blk.18.ssm_conv1d.weight +create_tensor: loading tensor blk.18.ssm_dt.bias +create_tensor: loading tensor blk.18.ssm_a +create_tensor: loading tensor blk.18.ssm_ba.weight +create_tensor: loading tensor blk.18.ssm_norm.weight +create_tensor: loading tensor blk.18.ssm_out.weight +create_tensor: loading tensor blk.18.ffn_gate_inp.weight +create_tensor: loading tensor blk.18.ffn_down_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_exps.weight +create_tensor: loading tensor blk.18.ffn_up_exps.weight +create_tensor: loading tensor blk.18.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.18.ffn_gate_shexp.weight +create_tensor: loading tensor blk.18.ffn_up_shexp.weight +create_tensor: loading tensor blk.18.ffn_down_shexp.weight +create_tensor: loading tensor blk.19.attn_norm.weight +create_tensor: loading tensor blk.19.post_attention_norm.weight +create_tensor: loading tensor blk.19.attn_q.weight +create_tensor: loading tensor blk.19.attn_k.weight +create_tensor: loading tensor blk.19.attn_v.weight +create_tensor: loading tensor blk.19.attn_output.weight +create_tensor: loading tensor blk.19.attn_q_norm.weight +create_tensor: loading tensor blk.19.attn_k_norm.weight +create_tensor: loading tensor blk.19.ffn_gate_inp.weight +create_tensor: loading tensor blk.19.ffn_down_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_exps.weight +create_tensor: loading tensor blk.19.ffn_up_exps.weight +create_tensor: loading tensor blk.19.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.19.ffn_gate_shexp.weight +create_tensor: loading tensor blk.19.ffn_up_shexp.weight +create_tensor: loading tensor blk.19.ffn_down_shexp.weight +create_tensor: loading tensor blk.20.attn_norm.weight +create_tensor: loading tensor blk.20.post_attention_norm.weight +create_tensor: loading tensor blk.20.attn_qkv.weight +create_tensor: loading tensor blk.20.attn_gate.weight +create_tensor: loading tensor blk.20.ssm_conv1d.weight +create_tensor: loading tensor blk.20.ssm_dt.bias +create_tensor: loading tensor blk.20.ssm_a +create_tensor: loading tensor blk.20.ssm_ba.weight +create_tensor: loading tensor blk.20.ssm_norm.weight +create_tensor: loading tensor blk.20.ssm_out.weight +create_tensor: loading tensor blk.20.ffn_gate_inp.weight +create_tensor: loading tensor blk.20.ffn_down_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_exps.weight +create_tensor: loading tensor blk.20.ffn_up_exps.weight +create_tensor: loading tensor blk.20.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.20.ffn_gate_shexp.weight +create_tensor: loading tensor blk.20.ffn_up_shexp.weight +create_tensor: loading tensor blk.20.ffn_down_shexp.weight +create_tensor: loading tensor blk.21.attn_norm.weight +create_tensor: loading tensor blk.21.post_attention_norm.weight +create_tensor: loading tensor blk.21.attn_qkv.weight +create_tensor: loading tensor blk.21.attn_gate.weight +create_tensor: loading tensor blk.21.ssm_conv1d.weight +create_tensor: loading tensor blk.21.ssm_dt.bias +create_tensor: loading tensor blk.21.ssm_a +create_tensor: loading tensor blk.21.ssm_ba.weight +create_tensor: loading tensor blk.21.ssm_norm.weight +create_tensor: loading tensor blk.21.ssm_out.weight +create_tensor: loading tensor blk.21.ffn_gate_inp.weight +create_tensor: loading tensor blk.21.ffn_down_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_exps.weight +create_tensor: loading tensor blk.21.ffn_up_exps.weight +create_tensor: loading tensor blk.21.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.21.ffn_gate_shexp.weight +create_tensor: loading tensor blk.21.ffn_up_shexp.weight +create_tensor: loading tensor blk.21.ffn_down_shexp.weight +create_tensor: loading tensor blk.22.attn_norm.weight +create_tensor: loading tensor blk.22.post_attention_norm.weight +create_tensor: loading tensor blk.22.attn_qkv.weight +create_tensor: loading tensor blk.22.attn_gate.weight +create_tensor: loading tensor blk.22.ssm_conv1d.weight +create_tensor: loading tensor blk.22.ssm_dt.bias +create_tensor: loading tensor blk.22.ssm_a +create_tensor: loading tensor blk.22.ssm_ba.weight +create_tensor: loading tensor blk.22.ssm_norm.weight +create_tensor: loading tensor blk.22.ssm_out.weight +create_tensor: loading tensor blk.22.ffn_gate_inp.weight +create_tensor: loading tensor blk.22.ffn_down_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_exps.weight +create_tensor: loading tensor blk.22.ffn_up_exps.weight +create_tensor: loading tensor blk.22.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.22.ffn_gate_shexp.weight +create_tensor: loading tensor blk.22.ffn_up_shexp.weight +create_tensor: loading tensor blk.22.ffn_down_shexp.weight +create_tensor: loading tensor blk.23.attn_norm.weight +create_tensor: loading tensor blk.23.post_attention_norm.weight +create_tensor: loading tensor blk.23.attn_q.weight +create_tensor: loading tensor blk.23.attn_k.weight +create_tensor: loading tensor blk.23.attn_v.weight +create_tensor: loading tensor blk.23.attn_output.weight +create_tensor: loading tensor blk.23.attn_q_norm.weight +create_tensor: loading tensor blk.23.attn_k_norm.weight +create_tensor: loading tensor blk.23.ffn_gate_inp.weight +tensor blk.23.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_exps.weight +create_tensor: loading tensor blk.23.ffn_up_exps.weight +create_tensor: loading tensor blk.23.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.23.ffn_gate_shexp.weight +create_tensor: loading tensor blk.23.ffn_up_shexp.weight +tensor blk.23.ffn_down_shexp.weight (0 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.23.ffn_down_shexp.weight +create_tensor: loading tensor blk.24.attn_norm.weight +create_tensor: loading tensor blk.24.post_attention_norm.weight +create_tensor: loading tensor blk.24.attn_qkv.weight +create_tensor: loading tensor blk.24.attn_gate.weight +create_tensor: loading tensor blk.24.ssm_conv1d.weight +create_tensor: loading tensor blk.24.ssm_dt.bias +create_tensor: loading tensor blk.24.ssm_a +create_tensor: loading tensor blk.24.ssm_ba.weight +create_tensor: loading tensor blk.24.ssm_norm.weight +create_tensor: loading tensor blk.24.ssm_out.weight +create_tensor: loading tensor blk.24.ffn_gate_inp.weight +tensor blk.24.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_down_exps.weight +tensor blk.24.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_gate_exps.weight +tensor blk.24.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.24.ffn_up_exps.weight +create_tensor: loading tensor blk.24.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.24.ffn_gate_shexp.weight +create_tensor: loading tensor blk.24.ffn_up_shexp.weight +create_tensor: loading tensor blk.24.ffn_down_shexp.weight +create_tensor: loading tensor blk.25.attn_norm.weight +create_tensor: loading tensor blk.25.post_attention_norm.weight +create_tensor: loading tensor blk.25.attn_qkv.weight +create_tensor: loading tensor blk.25.attn_gate.weight +create_tensor: loading tensor blk.25.ssm_conv1d.weight +create_tensor: loading tensor blk.25.ssm_dt.bias +create_tensor: loading tensor blk.25.ssm_a +create_tensor: loading tensor blk.25.ssm_ba.weight +create_tensor: loading tensor blk.25.ssm_norm.weight +create_tensor: loading tensor blk.25.ssm_out.weight +create_tensor: loading tensor blk.25.ffn_gate_inp.weight +tensor blk.25.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_down_exps.weight +tensor blk.25.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_gate_exps.weight +tensor blk.25.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.25.ffn_up_exps.weight +create_tensor: loading tensor blk.25.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.25.ffn_gate_shexp.weight +create_tensor: loading tensor blk.25.ffn_up_shexp.weight +create_tensor: loading tensor blk.25.ffn_down_shexp.weight +create_tensor: loading tensor blk.26.attn_norm.weight +create_tensor: loading tensor blk.26.post_attention_norm.weight +create_tensor: loading tensor blk.26.attn_qkv.weight +create_tensor: loading tensor blk.26.attn_gate.weight +create_tensor: loading tensor blk.26.ssm_conv1d.weight +create_tensor: loading tensor blk.26.ssm_dt.bias +create_tensor: loading tensor blk.26.ssm_a +create_tensor: loading tensor blk.26.ssm_ba.weight +create_tensor: loading tensor blk.26.ssm_norm.weight +create_tensor: loading tensor blk.26.ssm_out.weight +create_tensor: loading tensor blk.26.ffn_gate_inp.weight +tensor blk.26.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_down_exps.weight +tensor blk.26.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_gate_exps.weight +tensor blk.26.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.26.ffn_up_exps.weight +create_tensor: loading tensor blk.26.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.26.ffn_gate_shexp.weight +create_tensor: loading tensor blk.26.ffn_up_shexp.weight +create_tensor: loading tensor blk.26.ffn_down_shexp.weight +create_tensor: loading tensor blk.27.attn_norm.weight +create_tensor: loading tensor blk.27.post_attention_norm.weight +create_tensor: loading tensor blk.27.attn_q.weight +create_tensor: loading tensor blk.27.attn_k.weight +create_tensor: loading tensor blk.27.attn_v.weight +create_tensor: loading tensor blk.27.attn_output.weight +create_tensor: loading tensor blk.27.attn_q_norm.weight +create_tensor: loading tensor blk.27.attn_k_norm.weight +create_tensor: loading tensor blk.27.ffn_gate_inp.weight +tensor blk.27.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_down_exps.weight +tensor blk.27.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_gate_exps.weight +tensor blk.27.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.27.ffn_up_exps.weight +create_tensor: loading tensor blk.27.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.27.ffn_gate_shexp.weight +create_tensor: loading tensor blk.27.ffn_up_shexp.weight +create_tensor: loading tensor blk.27.ffn_down_shexp.weight +create_tensor: loading tensor blk.28.attn_norm.weight +create_tensor: loading tensor blk.28.post_attention_norm.weight +create_tensor: loading tensor blk.28.attn_qkv.weight +create_tensor: loading tensor blk.28.attn_gate.weight +create_tensor: loading tensor blk.28.ssm_conv1d.weight +create_tensor: loading tensor blk.28.ssm_dt.bias +create_tensor: loading tensor blk.28.ssm_a +create_tensor: loading tensor blk.28.ssm_ba.weight +create_tensor: loading tensor blk.28.ssm_norm.weight +create_tensor: loading tensor blk.28.ssm_out.weight +create_tensor: loading tensor blk.28.ffn_gate_inp.weight +tensor blk.28.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_down_exps.weight +tensor blk.28.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_gate_exps.weight +tensor blk.28.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.28.ffn_up_exps.weight +create_tensor: loading tensor blk.28.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.28.ffn_gate_shexp.weight +create_tensor: loading tensor blk.28.ffn_up_shexp.weight +create_tensor: loading tensor blk.28.ffn_down_shexp.weight +create_tensor: loading tensor blk.29.attn_norm.weight +create_tensor: loading tensor blk.29.post_attention_norm.weight +create_tensor: loading tensor blk.29.attn_qkv.weight +create_tensor: loading tensor blk.29.attn_gate.weight +create_tensor: loading tensor blk.29.ssm_conv1d.weight +create_tensor: loading tensor blk.29.ssm_dt.bias +create_tensor: loading tensor blk.29.ssm_a +create_tensor: loading tensor blk.29.ssm_ba.weight +create_tensor: loading tensor blk.29.ssm_norm.weight +create_tensor: loading tensor blk.29.ssm_out.weight +create_tensor: loading tensor blk.29.ffn_gate_inp.weight +tensor blk.29.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_down_exps.weight +tensor blk.29.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_gate_exps.weight +tensor blk.29.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.29.ffn_up_exps.weight +create_tensor: loading tensor blk.29.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.29.ffn_gate_shexp.weight +create_tensor: loading tensor blk.29.ffn_up_shexp.weight +create_tensor: loading tensor blk.29.ffn_down_shexp.weight +create_tensor: loading tensor blk.30.attn_norm.weight +create_tensor: loading tensor blk.30.post_attention_norm.weight +create_tensor: loading tensor blk.30.attn_qkv.weight +create_tensor: loading tensor blk.30.attn_gate.weight +create_tensor: loading tensor blk.30.ssm_conv1d.weight +create_tensor: loading tensor blk.30.ssm_dt.bias +create_tensor: loading tensor blk.30.ssm_a +create_tensor: loading tensor blk.30.ssm_ba.weight +create_tensor: loading tensor blk.30.ssm_norm.weight +create_tensor: loading tensor blk.30.ssm_out.weight +create_tensor: loading tensor blk.30.ffn_gate_inp.weight +tensor blk.30.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_down_exps.weight +tensor blk.30.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_gate_exps.weight +tensor blk.30.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.30.ffn_up_exps.weight +create_tensor: loading tensor blk.30.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.30.ffn_gate_shexp.weight +create_tensor: loading tensor blk.30.ffn_up_shexp.weight +create_tensor: loading tensor blk.30.ffn_down_shexp.weight +create_tensor: loading tensor blk.31.attn_norm.weight +create_tensor: loading tensor blk.31.post_attention_norm.weight +create_tensor: loading tensor blk.31.attn_q.weight +create_tensor: loading tensor blk.31.attn_k.weight +create_tensor: loading tensor blk.31.attn_v.weight +create_tensor: loading tensor blk.31.attn_output.weight +create_tensor: loading tensor blk.31.attn_q_norm.weight +create_tensor: loading tensor blk.31.attn_k_norm.weight +create_tensor: loading tensor blk.31.ffn_gate_inp.weight +tensor blk.31.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_down_exps.weight +tensor blk.31.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_gate_exps.weight +tensor blk.31.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.31.ffn_up_exps.weight +create_tensor: loading tensor blk.31.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.31.ffn_gate_shexp.weight +create_tensor: loading tensor blk.31.ffn_up_shexp.weight +create_tensor: loading tensor blk.31.ffn_down_shexp.weight +create_tensor: loading tensor blk.32.attn_norm.weight +create_tensor: loading tensor blk.32.post_attention_norm.weight +create_tensor: loading tensor blk.32.attn_qkv.weight +create_tensor: loading tensor blk.32.attn_gate.weight +create_tensor: loading tensor blk.32.ssm_conv1d.weight +create_tensor: loading tensor blk.32.ssm_dt.bias +create_tensor: loading tensor blk.32.ssm_a +create_tensor: loading tensor blk.32.ssm_ba.weight +create_tensor: loading tensor blk.32.ssm_norm.weight +create_tensor: loading tensor blk.32.ssm_out.weight +create_tensor: loading tensor blk.32.ffn_gate_inp.weight +tensor blk.32.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_down_exps.weight +tensor blk.32.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_gate_exps.weight +tensor blk.32.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.32.ffn_up_exps.weight +create_tensor: loading tensor blk.32.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.32.ffn_gate_shexp.weight +create_tensor: loading tensor blk.32.ffn_up_shexp.weight +create_tensor: loading tensor blk.32.ffn_down_shexp.weight +create_tensor: loading tensor blk.33.attn_norm.weight +create_tensor: loading tensor blk.33.post_attention_norm.weight +create_tensor: loading tensor blk.33.attn_qkv.weight +create_tensor: loading tensor blk.33.attn_gate.weight +create_tensor: loading tensor blk.33.ssm_conv1d.weight +create_tensor: loading tensor blk.33.ssm_dt.bias +create_tensor: loading tensor blk.33.ssm_a +create_tensor: loading tensor blk.33.ssm_ba.weight +create_tensor: loading tensor blk.33.ssm_norm.weight +create_tensor: loading tensor blk.33.ssm_out.weight +create_tensor: loading tensor blk.33.ffn_gate_inp.weight +tensor blk.33.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_down_exps.weight +tensor blk.33.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_gate_exps.weight +tensor blk.33.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.33.ffn_up_exps.weight +create_tensor: loading tensor blk.33.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.33.ffn_gate_shexp.weight +create_tensor: loading tensor blk.33.ffn_up_shexp.weight +create_tensor: loading tensor blk.33.ffn_down_shexp.weight +create_tensor: loading tensor blk.34.attn_norm.weight +create_tensor: loading tensor blk.34.post_attention_norm.weight +create_tensor: loading tensor blk.34.attn_qkv.weight +create_tensor: loading tensor blk.34.attn_gate.weight +create_tensor: loading tensor blk.34.ssm_conv1d.weight +create_tensor: loading tensor blk.34.ssm_dt.bias +create_tensor: loading tensor blk.34.ssm_a +create_tensor: loading tensor blk.34.ssm_ba.weight +create_tensor: loading tensor blk.34.ssm_norm.weight +create_tensor: loading tensor blk.34.ssm_out.weight +create_tensor: loading tensor blk.34.ffn_gate_inp.weight +tensor blk.34.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_down_exps.weight +tensor blk.34.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_gate_exps.weight +tensor blk.34.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.34.ffn_up_exps.weight +create_tensor: loading tensor blk.34.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.34.ffn_gate_shexp.weight +create_tensor: loading tensor blk.34.ffn_up_shexp.weight +create_tensor: loading tensor blk.34.ffn_down_shexp.weight +create_tensor: loading tensor blk.35.attn_norm.weight +create_tensor: loading tensor blk.35.post_attention_norm.weight +create_tensor: loading tensor blk.35.attn_q.weight +create_tensor: loading tensor blk.35.attn_k.weight +create_tensor: loading tensor blk.35.attn_v.weight +create_tensor: loading tensor blk.35.attn_output.weight +create_tensor: loading tensor blk.35.attn_q_norm.weight +create_tensor: loading tensor blk.35.attn_k_norm.weight +create_tensor: loading tensor blk.35.ffn_gate_inp.weight +tensor blk.35.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_down_exps.weight +tensor blk.35.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_gate_exps.weight +tensor blk.35.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.35.ffn_up_exps.weight +create_tensor: loading tensor blk.35.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.35.ffn_gate_shexp.weight +create_tensor: loading tensor blk.35.ffn_up_shexp.weight +create_tensor: loading tensor blk.35.ffn_down_shexp.weight +create_tensor: loading tensor blk.36.attn_norm.weight +create_tensor: loading tensor blk.36.post_attention_norm.weight +create_tensor: loading tensor blk.36.attn_qkv.weight +create_tensor: loading tensor blk.36.attn_gate.weight +create_tensor: loading tensor blk.36.ssm_conv1d.weight +create_tensor: loading tensor blk.36.ssm_dt.bias +create_tensor: loading tensor blk.36.ssm_a +create_tensor: loading tensor blk.36.ssm_ba.weight +create_tensor: loading tensor blk.36.ssm_norm.weight +create_tensor: loading tensor blk.36.ssm_out.weight +create_tensor: loading tensor blk.36.ffn_gate_inp.weight +tensor blk.36.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_down_exps.weight +tensor blk.36.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_gate_exps.weight +tensor blk.36.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.36.ffn_up_exps.weight +create_tensor: loading tensor blk.36.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.36.ffn_gate_shexp.weight +create_tensor: loading tensor blk.36.ffn_up_shexp.weight +create_tensor: loading tensor blk.36.ffn_down_shexp.weight +create_tensor: loading tensor blk.37.attn_norm.weight +create_tensor: loading tensor blk.37.post_attention_norm.weight +create_tensor: loading tensor blk.37.attn_qkv.weight +create_tensor: loading tensor blk.37.attn_gate.weight +create_tensor: loading tensor blk.37.ssm_conv1d.weight +create_tensor: loading tensor blk.37.ssm_dt.bias +create_tensor: loading tensor blk.37.ssm_a +create_tensor: loading tensor blk.37.ssm_ba.weight +create_tensor: loading tensor blk.37.ssm_norm.weight +create_tensor: loading tensor blk.37.ssm_out.weight +create_tensor: loading tensor blk.37.ffn_gate_inp.weight +tensor blk.37.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_down_exps.weight +tensor blk.37.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_gate_exps.weight +tensor blk.37.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.37.ffn_up_exps.weight +create_tensor: loading tensor blk.37.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.37.ffn_gate_shexp.weight +create_tensor: loading tensor blk.37.ffn_up_shexp.weight +create_tensor: loading tensor blk.37.ffn_down_shexp.weight +create_tensor: loading tensor blk.38.attn_norm.weight +create_tensor: loading tensor blk.38.post_attention_norm.weight +create_tensor: loading tensor blk.38.attn_qkv.weight +create_tensor: loading tensor blk.38.attn_gate.weight +create_tensor: loading tensor blk.38.ssm_conv1d.weight +create_tensor: loading tensor blk.38.ssm_dt.bias +create_tensor: loading tensor blk.38.ssm_a +create_tensor: loading tensor blk.38.ssm_ba.weight +create_tensor: loading tensor blk.38.ssm_norm.weight +create_tensor: loading tensor blk.38.ssm_out.weight +create_tensor: loading tensor blk.38.ffn_gate_inp.weight +tensor blk.38.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_down_exps.weight +tensor blk.38.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_gate_exps.weight +tensor blk.38.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.38.ffn_up_exps.weight +create_tensor: loading tensor blk.38.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.38.ffn_gate_shexp.weight +create_tensor: loading tensor blk.38.ffn_up_shexp.weight +create_tensor: loading tensor blk.38.ffn_down_shexp.weight +create_tensor: loading tensor blk.39.attn_norm.weight +create_tensor: loading tensor blk.39.post_attention_norm.weight +create_tensor: loading tensor blk.39.attn_q.weight +create_tensor: loading tensor blk.39.attn_k.weight +create_tensor: loading tensor blk.39.attn_v.weight +create_tensor: loading tensor blk.39.attn_output.weight +create_tensor: loading tensor blk.39.attn_q_norm.weight +create_tensor: loading tensor blk.39.attn_k_norm.weight +create_tensor: loading tensor blk.39.ffn_gate_inp.weight +tensor blk.39.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_down_exps.weight +tensor blk.39.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_gate_exps.weight +tensor blk.39.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.39.ffn_up_exps.weight +create_tensor: loading tensor blk.39.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.39.ffn_gate_shexp.weight +create_tensor: loading tensor blk.39.ffn_up_shexp.weight +create_tensor: loading tensor blk.39.ffn_down_shexp.weight +create_tensor: loading tensor blk.40.attn_norm.weight +create_tensor: loading tensor blk.40.post_attention_norm.weight +create_tensor: loading tensor blk.40.attn_qkv.weight +create_tensor: loading tensor blk.40.attn_gate.weight +create_tensor: loading tensor blk.40.ssm_conv1d.weight +create_tensor: loading tensor blk.40.ssm_dt.bias +create_tensor: loading tensor blk.40.ssm_a +create_tensor: loading tensor blk.40.ssm_ba.weight +create_tensor: loading tensor blk.40.ssm_norm.weight +create_tensor: loading tensor blk.40.ssm_out.weight +create_tensor: loading tensor blk.40.ffn_gate_inp.weight +tensor blk.40.ffn_down_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_down_exps.weight +tensor blk.40.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_gate_exps.weight +tensor blk.40.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.40.ffn_up_exps.weight +create_tensor: loading tensor blk.40.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.40.ffn_gate_shexp.weight +create_tensor: loading tensor blk.40.ffn_up_shexp.weight +create_tensor: loading tensor blk.40.ffn_down_shexp.weight +create_tensor: loading tensor blk.41.attn_norm.weight +create_tensor: loading tensor blk.41.post_attention_norm.weight +create_tensor: loading tensor blk.41.attn_qkv.weight +create_tensor: loading tensor blk.41.attn_gate.weight +create_tensor: loading tensor blk.41.ssm_conv1d.weight +create_tensor: loading tensor blk.41.ssm_dt.bias +create_tensor: loading tensor blk.41.ssm_a +create_tensor: loading tensor blk.41.ssm_ba.weight +create_tensor: loading tensor blk.41.ssm_norm.weight +create_tensor: loading tensor blk.41.ssm_out.weight +create_tensor: loading tensor blk.41.ffn_gate_inp.weight +tensor blk.41.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_down_exps.weight +tensor blk.41.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_gate_exps.weight +tensor blk.41.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.41.ffn_up_exps.weight +create_tensor: loading tensor blk.41.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.41.ffn_gate_shexp.weight +create_tensor: loading tensor blk.41.ffn_up_shexp.weight +create_tensor: loading tensor blk.41.ffn_down_shexp.weight +create_tensor: loading tensor blk.42.attn_norm.weight +create_tensor: loading tensor blk.42.post_attention_norm.weight +create_tensor: loading tensor blk.42.attn_qkv.weight +create_tensor: loading tensor blk.42.attn_gate.weight +create_tensor: loading tensor blk.42.ssm_conv1d.weight +create_tensor: loading tensor blk.42.ssm_dt.bias +create_tensor: loading tensor blk.42.ssm_a +create_tensor: loading tensor blk.42.ssm_ba.weight +create_tensor: loading tensor blk.42.ssm_norm.weight +create_tensor: loading tensor blk.42.ssm_out.weight +create_tensor: loading tensor blk.42.ffn_gate_inp.weight +tensor blk.42.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_down_exps.weight +tensor blk.42.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_gate_exps.weight +tensor blk.42.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.42.ffn_up_exps.weight +create_tensor: loading tensor blk.42.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.42.ffn_gate_shexp.weight +create_tensor: loading tensor blk.42.ffn_up_shexp.weight +create_tensor: loading tensor blk.42.ffn_down_shexp.weight +create_tensor: loading tensor blk.43.attn_norm.weight +create_tensor: loading tensor blk.43.post_attention_norm.weight +create_tensor: loading tensor blk.43.attn_q.weight +create_tensor: loading tensor blk.43.attn_k.weight +create_tensor: loading tensor blk.43.attn_v.weight +create_tensor: loading tensor blk.43.attn_output.weight +create_tensor: loading tensor blk.43.attn_q_norm.weight +create_tensor: loading tensor blk.43.attn_k_norm.weight +create_tensor: loading tensor blk.43.ffn_gate_inp.weight +tensor blk.43.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_down_exps.weight +tensor blk.43.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_gate_exps.weight +tensor blk.43.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.43.ffn_up_exps.weight +create_tensor: loading tensor blk.43.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.43.ffn_gate_shexp.weight +create_tensor: loading tensor blk.43.ffn_up_shexp.weight +create_tensor: loading tensor blk.43.ffn_down_shexp.weight +create_tensor: loading tensor blk.44.attn_norm.weight +create_tensor: loading tensor blk.44.post_attention_norm.weight +create_tensor: loading tensor blk.44.attn_qkv.weight +create_tensor: loading tensor blk.44.attn_gate.weight +create_tensor: loading tensor blk.44.ssm_conv1d.weight +create_tensor: loading tensor blk.44.ssm_dt.bias +create_tensor: loading tensor blk.44.ssm_a +create_tensor: loading tensor blk.44.ssm_ba.weight +create_tensor: loading tensor blk.44.ssm_norm.weight +create_tensor: loading tensor blk.44.ssm_out.weight +create_tensor: loading tensor blk.44.ffn_gate_inp.weight +tensor blk.44.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_down_exps.weight +tensor blk.44.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_gate_exps.weight +tensor blk.44.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.44.ffn_up_exps.weight +create_tensor: loading tensor blk.44.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.44.ffn_gate_shexp.weight +create_tensor: loading tensor blk.44.ffn_up_shexp.weight +create_tensor: loading tensor blk.44.ffn_down_shexp.weight +create_tensor: loading tensor blk.45.attn_norm.weight +create_tensor: loading tensor blk.45.post_attention_norm.weight +create_tensor: loading tensor blk.45.attn_qkv.weight +create_tensor: loading tensor blk.45.attn_gate.weight +create_tensor: loading tensor blk.45.ssm_conv1d.weight +create_tensor: loading tensor blk.45.ssm_dt.bias +create_tensor: loading tensor blk.45.ssm_a +create_tensor: loading tensor blk.45.ssm_ba.weight +create_tensor: loading tensor blk.45.ssm_norm.weight +create_tensor: loading tensor blk.45.ssm_out.weight +create_tensor: loading tensor blk.45.ffn_gate_inp.weight +tensor blk.45.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_down_exps.weight +tensor blk.45.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_gate_exps.weight +tensor blk.45.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.45.ffn_up_exps.weight +create_tensor: loading tensor blk.45.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.45.ffn_gate_shexp.weight +create_tensor: loading tensor blk.45.ffn_up_shexp.weight +create_tensor: loading tensor blk.45.ffn_down_shexp.weight +create_tensor: loading tensor blk.46.attn_norm.weight +create_tensor: loading tensor blk.46.post_attention_norm.weight +create_tensor: loading tensor blk.46.attn_qkv.weight +create_tensor: loading tensor blk.46.attn_gate.weight +create_tensor: loading tensor blk.46.ssm_conv1d.weight +create_tensor: loading tensor blk.46.ssm_dt.bias +create_tensor: loading tensor blk.46.ssm_a +create_tensor: loading tensor blk.46.ssm_ba.weight +create_tensor: loading tensor blk.46.ssm_norm.weight +create_tensor: loading tensor blk.46.ssm_out.weight +create_tensor: loading tensor blk.46.ffn_gate_inp.weight +tensor blk.46.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_down_exps.weight +tensor blk.46.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_gate_exps.weight +tensor blk.46.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.46.ffn_up_exps.weight +create_tensor: loading tensor blk.46.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.46.ffn_gate_shexp.weight +create_tensor: loading tensor blk.46.ffn_up_shexp.weight +create_tensor: loading tensor blk.46.ffn_down_shexp.weight +create_tensor: loading tensor blk.47.attn_norm.weight +create_tensor: loading tensor blk.47.post_attention_norm.weight +create_tensor: loading tensor blk.47.attn_q.weight +create_tensor: loading tensor blk.47.attn_k.weight +create_tensor: loading tensor blk.47.attn_v.weight +create_tensor: loading tensor blk.47.attn_output.weight +create_tensor: loading tensor blk.47.attn_q_norm.weight +create_tensor: loading tensor blk.47.attn_k_norm.weight +create_tensor: loading tensor blk.47.ffn_gate_inp.weight +tensor blk.47.ffn_down_exps.weight (420 MiB q6_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_down_exps.weight +tensor blk.47.ffn_gate_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_gate_exps.weight +tensor blk.47.ffn_up_exps.weight (352 MiB q5_K) buffer type overridden to ROCm_Host +create_tensor: loading tensor blk.47.ffn_up_exps.weight +create_tensor: loading tensor blk.47.ffn_gate_inp_shexp.weight +create_tensor: loading tensor blk.47.ffn_gate_shexp.weight +create_tensor: loading tensor blk.47.ffn_up_shexp.weight +create_tensor: loading tensor blk.47.ffn_down_shexp.weight +done_getting_tensors: tensor 'token_embd.weight' (q5_K) (and 74 others) cannot be used with preferred buffer type ROCm_Host, using CPU instead +load_tensors: offloading output layer to GPU +load_tensors: offloading 47 repeating layers to GPU +load_tensors: offloaded 49/49 layers to GPU +load_tensors: CPU model buffer size = 204.02 MiB +load_tensors: ROCm0 model buffer size = 27423.82 MiB +load_tensors: ROCm_Host model buffer size = 26580.82 MiB +load_all_data: no device found for buffer type CPU for async uploads +load_all_data: using async uploads for device ROCm0, buffer type ROCm0, backend ROCm0 +..................................................load_all_data: buffer type ROCm_Host is not the default buffer type for device ROCm0 for async uploads +.................................................. +common_init_result: added logit bias = -inf +common_init_result: added <|endoftext|> logit bias = -inf +common_init_result: added <|im_end|> logit bias = -inf +common_init_result: added <|fim_pad|> logit bias = -inf +common_init_result: added <|repo_name|> logit bias = -inf +common_init_result: added <|file_sep|> logit bias = -inf +llama_context: constructing llama_context +llama_context: n_seq_max = 1 +llama_context: n_ctx = 131072 +llama_context: n_ctx_seq = 131072 +llama_context: n_batch = 2048 +llama_context: n_ubatch = 512 +llama_context: causal_attn = 1 +llama_context: flash_attn = enabled +llama_context: kv_unified = false +llama_context: freq_base = 5000000.0 +llama_context: freq_scale = 1 +llama_context: n_ctx_seq (131072) < n_ctx_train (262144) -- the full capacity of the model will not be utilized +set_abort_callback: call +llama_context: ROCm_Host output buffer size = 0.58 MiB +llama_kv_cache: layer 0: filtered +llama_kv_cache: layer 1: filtered +llama_kv_cache: layer 2: filtered +llama_kv_cache: layer 3: dev = ROCm0 +llama_kv_cache: layer 4: filtered +llama_kv_cache: layer 5: filtered +llama_kv_cache: layer 6: filtered +llama_kv_cache: layer 7: dev = ROCm0 +llama_kv_cache: layer 8: filtered +llama_kv_cache: layer 9: filtered +llama_kv_cache: layer 10: filtered +llama_kv_cache: layer 11: dev = ROCm0 +llama_kv_cache: layer 12: filtered +llama_kv_cache: layer 13: filtered +llama_kv_cache: layer 14: filtered +llama_kv_cache: layer 15: dev = ROCm0 +llama_kv_cache: layer 16: filtered +llama_kv_cache: layer 17: filtered +llama_kv_cache: layer 18: filtered +llama_kv_cache: layer 19: dev = ROCm0 +llama_kv_cache: layer 20: filtered +llama_kv_cache: layer 21: filtered +llama_kv_cache: layer 22: filtered +llama_kv_cache: layer 23: dev = ROCm0 +llama_kv_cache: layer 24: filtered +llama_kv_cache: layer 25: filtered +llama_kv_cache: layer 26: filtered +llama_kv_cache: layer 27: dev = ROCm0 +llama_kv_cache: layer 28: filtered +llama_kv_cache: layer 29: filtered +llama_kv_cache: layer 30: filtered +llama_kv_cache: layer 31: dev = ROCm0 +llama_kv_cache: layer 32: filtered +llama_kv_cache: layer 33: filtered +llama_kv_cache: layer 34: filtered +llama_kv_cache: layer 35: dev = ROCm0 +llama_kv_cache: layer 36: filtered +llama_kv_cache: layer 37: filtered +llama_kv_cache: layer 38: filtered +llama_kv_cache: layer 39: dev = ROCm0 +llama_kv_cache: layer 40: filtered +llama_kv_cache: layer 41: filtered +llama_kv_cache: layer 42: filtered +llama_kv_cache: layer 43: dev = ROCm0 +llama_kv_cache: layer 44: filtered +llama_kv_cache: layer 45: filtered +llama_kv_cache: layer 46: filtered +llama_kv_cache: layer 47: dev = ROCm0 +llama_kv_cache: ROCm0 KV buffer size = 3072.00 MiB +llama_kv_cache: size = 3072.00 MiB (131072 cells, 12 layers, 1/1 seqs), K (f16): 1536.00 MiB, V (f16): 1536.00 MiB +llama_memory_recurrent, layer 0: dev = ROCm0 +llama_memory_recurrent, layer 1: dev = ROCm0 +llama_memory_recurrent, layer 2: dev = ROCm0 +llama_memory_recurrent: layer 3: skipped +llama_memory_recurrent, layer 4: dev = ROCm0 +llama_memory_recurrent, layer 5: dev = ROCm0 +llama_memory_recurrent, layer 6: dev = ROCm0 +llama_memory_recurrent: layer 7: skipped +llama_memory_recurrent, layer 8: dev = ROCm0 +llama_memory_recurrent, layer 9: dev = ROCm0 +llama_memory_recurrent, layer 10: dev = ROCm0 +llama_memory_recurrent: layer 11: skipped +llama_memory_recurrent, layer 12: dev = ROCm0 +llama_memory_recurrent, layer 13: dev = ROCm0 +llama_memory_recurrent, layer 14: dev = ROCm0 +llama_memory_recurrent: layer 15: skipped +llama_memory_recurrent, layer 16: dev = ROCm0 +llama_memory_recurrent, layer 17: dev = ROCm0 +llama_memory_recurrent, layer 18: dev = ROCm0 +llama_memory_recurrent: layer 19: skipped +llama_memory_recurrent, layer 20: dev = ROCm0 +llama_memory_recurrent, layer 21: dev = ROCm0 +llama_memory_recurrent, layer 22: dev = ROCm0 +llama_memory_recurrent: layer 23: skipped +llama_memory_recurrent, layer 24: dev = ROCm0 +llama_memory_recurrent, layer 25: dev = ROCm0 +llama_memory_recurrent, layer 26: dev = ROCm0 +llama_memory_recurrent: layer 27: skipped +llama_memory_recurrent, layer 28: dev = ROCm0 +llama_memory_recurrent, layer 29: dev = ROCm0 +llama_memory_recurrent, layer 30: dev = ROCm0 +llama_memory_recurrent: layer 31: skipped +llama_memory_recurrent, layer 32: dev = ROCm0 +llama_memory_recurrent, layer 33: dev = ROCm0 +llama_memory_recurrent, layer 34: dev = ROCm0 +llama_memory_recurrent: layer 35: skipped +llama_memory_recurrent, layer 36: dev = ROCm0 +llama_memory_recurrent, layer 37: dev = ROCm0 +llama_memory_recurrent, layer 38: dev = ROCm0 +llama_memory_recurrent: layer 39: skipped +llama_memory_recurrent, layer 40: dev = ROCm0 +llama_memory_recurrent, layer 41: dev = ROCm0 +llama_memory_recurrent, layer 42: dev = ROCm0 +llama_memory_recurrent: layer 43: skipped +llama_memory_recurrent, layer 44: dev = ROCm0 +llama_memory_recurrent, layer 45: dev = ROCm0 +llama_memory_recurrent, layer 46: dev = ROCm0 +llama_memory_recurrent: layer 47: skipped +llama_memory_recurrent: ROCm0 RS buffer size = 75.38 MiB +llama_memory_recurrent: size = 75.38 MiB ( 1 cells, 48 layers, 1 seqs), R (f32): 3.38 MiB, S (f32): 72.00 MiB +llama_context: enumerating backends +llama_context: backend_ptrs.size() = 2 +sched_reserve: reserving ... +sched_reserve: max_nodes = 26976 +sched_reserve: reserving full memory module +sched_reserve: worst-case: n_tokens = 512, n_seqs = 1, n_outputs = 1 +sched_reserve: resolving fused Gated Delta Net support: +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +sched_reserve: fused Gated Delta Net (autoregressive) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 16, n_seqs = 1, n_outputs = 16 +sched_reserve: fused Gated Delta Net (chunked) enabled +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +graph_reserve: reserving a graph for ubatch with n_tokens = 1, n_seqs = 1, n_outputs = 1 +graph_reserve: reserving a graph for ubatch with n_tokens = 512, n_seqs = 1, n_outputs = 512 +sched_reserve: ROCm0 compute buffer size = 840.01 MiB +sched_reserve: ROCm_Host compute buffer size = 264.01 MiB +sched_reserve: graph nodes = 5013 +sched_reserve: graph splits = 76 (with bs=512), 54 (with bs=1) +sched_reserve: reserve took 274.49 ms, sched copies = 1 +set_adapters_lora: adapters = (nil) +adapters_lora_are_same: adapters = (nil) +common_init_from_params: warming up the model with an empty run - please wait ... (--no-warmup to disable) +set_warmup: value = 1 +set_warmup: value = 0 +srv load_model: initializing slots, n_slots = 1 +common_speculative_is_compat: the target context does not support partial sequence removal +srv load_model: speculative decoding not supported by this context +slot load_model: id 0 | task -1 | new slot, n_ctx = 131072 +slot reset: id 0 | task -1 | +srv load_model: prompt cache is enabled, size limit: 8192 MiB +srv load_model: use `--cache-ram 0` to disable the prompt cache +srv load_model: for more info see https://github.com/ggml-org/llama.cpp/pull/16391 +Using differential autoparser +=== Starting differential analysis === +Phase 1: Reasoning analysis +Phase 2: Content analysis +Phase 3: Tool call analysis +Phase 4: Argument analysis + +--- Reasoning & Content Structure --- +reasoning_mode: NONE +reasoning_start: '' +reasoning_end: '' +content_mode: PLAIN +content_start: '' +content_end: '' + +--- Tool Call Structure --- +tool_mode: TAG_WITH_TAGGED +supports_tools: true +supports_parallel_calls: true +tool_section_start: '' +tool_section_end: '' +per_call_start: ' +' +per_call_end: '' +func_name_prefix: ' +' +func_close: ' +' +python_dict_format: false +arg_name_prefix: ' +' +arg_value_prefix: '' +arg_value_suffix: ' +' +name_field: 'name' +args_field: 'arguments' +id_field: '' +gen_id_field: '' +parameter_order: '' +=== Differential analysis complete === +init: chat template, example_format: '<|im_start|>system +You are a helpful assistant<|im_end|> +<|im_start|>user +Hello<|im_end|> +<|im_start|>assistant +Hi there<|im_end|> +<|im_start|>user +How are you?<|im_end|> +<|im_start|>assistant +' +Using differential autoparser +=== Starting differential analysis === +Phase 1: Reasoning analysis +Phase 2: Content analysis +Phase 3: Tool call analysis +Phase 4: Argument analysis + +--- Reasoning & Content Structure --- +reasoning_mode: NONE +reasoning_start: '' +reasoning_end: '' +content_mode: PLAIN +content_start: '' +content_end: '' + +--- Tool Call Structure --- +tool_mode: TAG_WITH_TAGGED +supports_tools: true +supports_parallel_calls: true +tool_section_start: '' +tool_section_end: '' +per_call_start: ' +' +per_call_end: '' +func_name_prefix: ' +' +func_close: ' +' +python_dict_format: false +arg_name_prefix: ' +' +arg_value_prefix: '' +arg_value_suffix: ' +' +name_field: 'name' +args_field: 'arguments' +id_field: '' +gen_id_field: '' +parameter_order: '' +=== Differential analysis complete === +srv init: init: chat template, thinking = 0 +main: model loaded +main: server is listening on http://0.0.0.0:8001 +main: starting the main loop... +que start_loop: processing new tasks +que start_loop: update slots +srv update_slots: all slots are idle +que start_loop: waiting for new tasks diff --git a/src/command/dispatch.rs b/src/command/dispatch.rs index 1495811..dd8d26a 100644 --- a/src/command/dispatch.rs +++ b/src/command/dispatch.rs @@ -163,50 +163,69 @@ pub fn dispatch(model: &mut Model, cmd: &Command) -> CommandResult { path, model_name, array_path, - } => import_json_headless(model, path, model_name.as_deref(), array_path.as_deref()), + } => import_headless(model, path, model_name.as_deref(), array_path.as_deref()), } } -fn import_json_headless( +fn import_headless( model: &mut Model, path: &str, model_name: Option<&str>, array_path: Option<&str>, ) -> CommandResult { - let content = match std::fs::read_to_string(path) { - Ok(c) => c, - Err(e) => return CommandResult::err(format!("Cannot read '{path}': {e}")), - }; - let value: serde_json::Value = match serde_json::from_str(&content) { - Ok(v) => v, - Err(e) => return CommandResult::err(format!("JSON parse error: {e}")), - }; + let is_csv = path.ends_with(".csv"); - let records = if let Some(ap) = array_path.filter(|s| !s.is_empty()) { - match extract_array_at_path(&value, ap) { - Some(arr) => arr.clone(), - None => return CommandResult::err(format!("No array at path '{ap}'")), + let records = if is_csv { + // Parse CSV file + match crate::import::csv_parser::parse_csv(path) { + Ok(recs) => recs, + Err(e) => return CommandResult::err(e.to_string()), } - } else if let Some(arr) = value.as_array() { - arr.clone() } else { - // Find first array - let paths = crate::import::analyzer::find_array_paths(&value); - if let Some(first) = paths.first() { - match extract_array_at_path(&value, first) { + // Parse JSON file + let content = match std::fs::read_to_string(path) { + Ok(c) => c, + Err(e) => return CommandResult::err(format!("Cannot read '{path}': {e}")), + }; + let value: serde_json::Value = match serde_json::from_str(&content) { + Ok(v) => v, + Err(e) => return CommandResult::err(format!("JSON parse error: {e}")), + }; + + if let Some(ap) = array_path.filter(|s| !s.is_empty()) { + match extract_array_at_path(&value, ap) { Some(arr) => arr.clone(), - None => return CommandResult::err("Could not extract records array"), + None => return CommandResult::err(format!("No array at path '{ap}'")), } + } else if let Some(arr) = value.as_array() { + arr.clone() } else { - return CommandResult::err("No array found in JSON"); + let paths = crate::import::analyzer::find_array_paths(&value); + if let Some(first) = paths.first() { + match extract_array_at_path(&value, first) { + Some(arr) => arr.clone(), + None => return CommandResult::err("Could not extract records array"), + } + } else { + return CommandResult::err("No array found in JSON"); + } } }; let proposals = analyze_records(&records); - // Auto-accept all and build via ImportPipeline + // Build via ImportPipeline + let raw = if is_csv { + serde_json::Value::Array(records.clone()) + } else { + // For JSON, we need the original parsed value + // Re-read and parse to get it (or pass it up from above) + serde_json::from_str(&std::fs::read_to_string(path).unwrap_or_default()) + .unwrap_or(serde_json::Value::Array(records.clone())) + }; + let pipeline = crate::import::wizard::ImportPipeline { - raw: value, + raw, array_paths: vec![], selected_path: array_path.unwrap_or("").to_string(), records, @@ -223,7 +242,7 @@ fn import_json_headless( match pipeline.build_model() { Ok(new_model) => { *model = new_model; - CommandResult::ok_msg("JSON imported successfully") + CommandResult::ok_msg("Imported successfully") } Err(e) => CommandResult::err(e.to_string()), } diff --git a/src/import/csv_parser.rs b/src/import/csv_parser.rs new file mode 100644 index 0000000..8342c61 --- /dev/null +++ b/src/import/csv_parser.rs @@ -0,0 +1,159 @@ +use anyhow::{Context, Result}; +use csv::ReaderBuilder; +use serde_json::Value; + +/// Parse a CSV file and return records as serde_json::Value array +pub fn parse_csv(path: &str) -> Result> { + let mut reader = ReaderBuilder::new() + .has_headers(true) + .flexible(true) + .trim(csv::Trim::All) + .from_path(path) + .with_context(|| format!("Failed to open CSV file: {path}"))?; + + // Detect if first row looks like headers (strings) or data (mixed) + let has_headers = reader.headers().is_ok(); + + let mut records = Vec::new(); + let mut headers = Vec::new(); + + if has_headers { + headers = reader + .headers() + .with_context(|| "Failed to read CSV headers")? + .iter() + .map(|s| s.to_string()) + .collect(); + } + + for result in reader.records() { + let record = result.with_context(|| "Failed to read CSV record")?; + let mut map = serde_json::Map::new(); + + for (i, field) in record.iter().enumerate() { + let json_value: Value = parse_csv_field(field); + if has_headers { + if let Some(header) = headers.get(i) { + map.insert(header.clone(), json_value); + } + } else { + map.insert(i.to_string(), json_value); + } + } + + if !map.is_empty() { + records.push(Value::Object(map)); + } + } + + Ok(records) +} + +fn parse_csv_field(field: &str) -> Value { + if field.is_empty() { + return Value::Null; + } + + // Try to parse as number (integer or float) + if let Ok(num) = field.parse::() { + return Value::Number(serde_json::Number::from(num)); + } + + if let Ok(num) = field.parse::() { + return Value::Number( + serde_json::Number::from_f64(num).unwrap_or(serde_json::Number::from(0)), + ); + } + + // Otherwise treat as string + Value::String(field.to_string()) +} + +#[cfg(test)] +mod tests { + use super::*; + use std::fs; + use tempfile::tempdir; + + fn create_temp_csv(content: &str) -> (String, tempfile::TempDir) { + let dir = tempdir().unwrap(); + let path = dir.path().join("test.csv"); + fs::write(&path, content).unwrap(); + (path.to_string_lossy().to_string(), dir) + } + + #[test] + fn parse_simple_csv() { + let (path, _dir) = create_temp_csv("Region,Product,Revenue\nEast,Shirts,1000\nWest,Shirts,800"); + let records = parse_csv(&path).unwrap(); + + assert_eq!(records.len(), 2); + assert_eq!(records[0]["Region"], Value::String("East".to_string())); + assert_eq!(records[0]["Product"], Value::String("Shirts".to_string())); + assert_eq!(records[0]["Revenue"], Value::Number(serde_json::Number::from(1000))); + } + + #[test] + fn parse_csv_with_floats() { + let (path, _dir) = + create_temp_csv("Region,Revenue,Cost\nEast,1000.50,600.25\nWest,800.75,500.00"); + let records = parse_csv(&path).unwrap(); + + assert_eq!(records.len(), 2); + assert!(records[0]["Revenue"].is_f64()); + assert_eq!(records[0]["Revenue"], Value::Number(serde_json::Number::from_f64(1000.50).unwrap())); + } + + #[test] + fn parse_csv_with_quoted_fields() { + let (path, _dir) = create_temp_csv("Product,Description,Price\n\"Shirts\",\"A nice shirt\",10.00"); + let records = parse_csv(&path).unwrap(); + + assert_eq!(records.len(), 1); + assert_eq!(records[0]["Product"], Value::String("Shirts".to_string())); + assert_eq!(records[0]["Description"], Value::String("A nice shirt".to_string())); + } + + #[test] + fn parse_csv_with_empty_values() { + let (path, _dir) = create_temp_csv("Region,Product,Revenue\nEast,,1000\nWest,Shirts,"); + let records = parse_csv(&path).unwrap(); + + assert_eq!(records.len(), 2); + assert_eq!(records[0]["Product"], Value::Null); + assert_eq!(records[1]["Revenue"], Value::Null); + } + + #[test] + fn parse_csv_mixed_types() { + let (path, _dir) = create_temp_csv( + "Name,Count,Price,Active\nWidget,5,9.99,true\nGadget,3,19.99,false", + ); + let records = parse_csv(&path).unwrap(); + + assert_eq!(records.len(), 2); + assert_eq!(records[0]["Name"], Value::String("Widget".to_string())); + assert_eq!(records[0]["Count"], Value::Number(serde_json::Number::from(5))); + assert!(records[0]["Price"].is_f64()); + assert_eq!(records[0]["Active"], Value::String("true".to_string())); + } + + #[test] + fn parse_checking_csv_format() { + // Simulates the format of /Users/edwlan/Downloads/Checking1.csv + let (path, _dir) = create_temp_csv( + "Date,Amount,Flag,CheckNo,Description\n\ + \"03/31/2026\",\"-50.00\",\"*\",\"\",\"VENMO PAYMENT 260331\"\n\ + \"03/31/2026\",\"-240.00\",\"*\",\"\",\"ROBINHOOD DEBITS XXXXX3795\"", + ); + let records = parse_csv(&path).unwrap(); + + assert_eq!(records.len(), 2); + assert_eq!(records[0]["Date"], Value::String("03/31/2026".to_string())); + assert_eq!(records[0]["Amount"], Value::Number(serde_json::Number::from_f64(-50.00).unwrap())); + assert_eq!(records[0]["Flag"], Value::String("*".to_string())); + assert_eq!(records[0]["CheckNo"], Value::Null); + assert_eq!(records[0]["Description"], Value::String("VENMO PAYMENT 260331".to_string())); + assert_eq!(records[1]["Amount"], Value::Number(serde_json::Number::from_f64(-240.00).unwrap())); + } +} diff --git a/src/import/mod.rs b/src/import/mod.rs index 91e28e4..9082358 100644 --- a/src/import/mod.rs +++ b/src/import/mod.rs @@ -1,2 +1,3 @@ pub mod analyzer; +pub mod csv_parser; pub mod wizard; diff --git a/src/main.rs b/src/main.rs index edb9ed6..c34a421 100644 --- a/src/main.rs +++ b/src/main.rs @@ -54,26 +54,40 @@ impl Runnable for CmdLineArgs { // Load or create model let model = get_initial_model(&self.file_path)?; - // Pre-TUI import: parse JSON and open wizard - let import_json = if let Some(ref path) = self.import_path { + // Pre-TUI import: parse JSON or CSV and open wizard + let import_value = if let Some(ref path) = self.import_path { match std::fs::read_to_string(path) { Err(e) => { eprintln!("Cannot read '{}': {e}", path.display()); return Ok(()); } - Ok(content) => match serde_json::from_str::(&content) { - Err(e) => { - eprintln!("JSON parse error: {e}"); - return Ok(()); + Ok(content) => { + if path.to_string_lossy().ends_with(".csv") { + // Parse CSV and wrap as JSON array + match crate::import::csv_parser::parse_csv(&path.to_string_lossy()) { + Ok(records) => Some(serde_json::Value::Array(records)), + Err(e) => { + eprintln!("CSV parse error: {e}"); + return Ok(()); + } + } + } else { + // Parse JSON + match serde_json::from_str::(&content) { + Err(e) => { + eprintln!("JSON parse error: {e}"); + return Ok(()); + } + Ok(json) => Some(json), + } } - Ok(json) => Some(json), - }, + } } } else { None }; - run_tui(model, self.file_path, import_json) + run_tui(model, self.file_path, import_value) } } @@ -130,7 +144,7 @@ impl Runnable for HelpArgs { println!("improvise — multi-dimensional data modeling TUI\n"); println!("USAGE:"); println!(" improvise [file.improv] Open or create a model"); - println!(" improvise --import data.json Import JSON then open TUI"); + println!(" improvise --import data.json Import JSON (or CSV) then open TUI"); println!(" improvise --cmd '{{...}}' Run a JSON command (headless, repeatable)"); println!(" improvise --script cmds.jsonl Run commands from file (headless)"); println!("\nTUI KEYS (vim-style):"); @@ -246,13 +260,13 @@ impl<'a> Drop for TuiContext<'a> { fn run_tui( model: Model, file_path: Option, - import_json: Option, + import_value: Option, ) -> Result<()> { let mut stdout = io::stdout(); let mut tui_context = TuiContext::enter(&mut stdout)?; let mut app = App::new(model, file_path); - if let Some(json) = import_json { + if let Some(json) = import_value { app.start_import_wizard(json); } @@ -518,7 +532,7 @@ fn draw_welcome(f: &mut Frame, area: Rect) { ), ("", Style::default()), ( - ":import Import a JSON file", + ":import Import JSON or CSV file", Style::default().fg(Color::Cyan), ), ( -- 2.51.0