package notebook import ( "bufio" "bytes" "io" "os" "path/filepath" "strings" "github.com/databricks/databricks-sdk-go/service/workspace" ) // Maximum length in bytes of the notebook header. const headerLength = 32 // readHeader reads the first N bytes from a file. func readHeader(path string) ([]byte, error) { f, err := os.Open(path) if err != nil { return nil, err } defer f.Close() // Scan header line with some padding. var buf = make([]byte, headerLength) n, err := f.Read([]byte(buf)) if err != nil && err != io.EOF { return nil, err } // Trim buffer to actual read bytes. return buf[:n], nil } // Detect returns whether the file at path is a Databricks notebook. // If it is, it returns the notebook language. func Detect(path string) (notebook bool, language workspace.Language, err error) { header := "" // Determine which header to expect based on filename extension. ext := strings.ToLower(filepath.Ext(path)) switch ext { case ".py": header = `# Databricks notebook source` language = workspace.LanguagePython case ".r": header = `# Databricks notebook source` language = workspace.LanguageR case ".scala": header = "// Databricks notebook source" language = workspace.LanguageScala case ".sql": header = "-- Databricks notebook source" language = workspace.LanguageSql default: return false, "", nil } buf, err := readHeader(path) if err != nil { return false, "", err } scanner := bufio.NewScanner(bytes.NewReader(buf)) scanner.Scan() if scanner.Text() != header { return false, "", nil } return true, language, nil }