印刻万物 TOP3DGS印刻万物TOP3DGS

stage 03

相机参数与现场操作

PublishedLast reviewed: 2026-05-08

Camera Settings & On-Site Operation

The 5 Iron Rules (Universal)

Regardless of device, these 5 rules are non-negotiable. Burn them into muscle memory before worrying about device-specific differences.

#RuleWhy
1M mode locks everything (aperture / shutter / ISO / WB / focal length / focus)The algorithm assumes "the same point has consistent color from different angles." Any auto-varying parameter breaks this assumption.
2One lens for the entire session, no zoomingZooming changes focal length = changes the intrinsic matrix. SfM will collapse.
3HDR always on3DGS suffers from detail loss in blown highlights / crushed shadows. HDR recovers both.
4Adjacent frame overlap ≥ 50% (70% recommended)SfM relies on feature point matching between adjacent frames to locate camera positions. Insufficient overlap = position unsolvable.
5Translate the camera, don't rotate in placeRotation produces no parallax, zero geometric information. The algorithm cannot find a triangulation baseline.

Lighting Condition Priority

3DGS loves uniform diffuse light — its worst enemy is "the specular highlight moved between angles." Inconsistent lighting directly causes color discontinuities across viewing angles in the trained model.

figure

Lighting conditionRatingReason
Overcast / cloudy★★★★★Most uniform diffuse light, no harsh directional light, virtually no shadow artifacts
Sunny · 2 hours after sunrise★★★★Soft light, stable color temp ~5200K, but avoid noon
Sunny · noon★★Overhead light creates harsh shadows, subject contrast >8:1, color distortion likely
Indoor · natural light★★★Turn off TV screens, spinning chandeliers — keep light stable
Indoor · 3-point softbox★★★★★Fully controllable, studio first choice. Unified 5600K, CRI ≥95
Rain / fogLens gets wet, atmospheric scattering too strong, geometric information lost

Rule of thumb: Overcast skies are God's gift to 3DGS shooters. If you can choose the date, always choose overcast.

Video Frame Extraction vs Photo Burst

The industry is split on this. Our judgment: beginners use video, professionals use photos.

DimensionVideo extraction (60fps+)Photo burst
OverlapAutomatic 70%+, no thinking requiredManual rhythm control needed
Per-frame resolutionSame CMOS: video is ~1/4 pixels of photo (4K=8MP vs photo 50MP)Full sensor resolution
Motion blurLikely (movement + lower shutter speed)Minimal (high shutter ≥1/125s)
Post color grading headroomAlmost none (H.264/H.265 compression)RAW gives 14-bit latitude
Best forBeginners, mobile devices, speedExperienced shooters, quality

Device-Specific Parameters

Below is an excerpt from our studio's capture manual, providing "power-on-and-go" parameter presets by device.

figure

Phone (iPhone / Android Flagship)

SettingValueNotes
ModeM / Pro modeiPhone's native camera won't work — use Halide / ProCamera
Lens1× main0.5× ultrawide only for small-space rescue
Resolution4032×3024 (4:3)Don't use 16:9 — loses ~25% vertical information
ISO100–400Don't exceed 800 in low light (noise destroys feature matching)
Shutter≥1/60s (handheld) / 1/30s (tripod)Prevent motion blur
HDRAlways oniPhone default on; Android needs manual enable
FocusLock (tap AE/AF Lock)Critical
White balanceManual 5500KNever Auto — Auto jumps between angles

iPhone note: The system camera's "Auto HDR" actually varies exposure strategy per frame, causing brightness inconsistency between adjacent frames. Use Halide's Manual HDR mode, or disable system HDR entirely and use fixed exposure.

Mirrorless / DSLR

SettingValueNotes
ModeM + RAWRAW provides 14-bit post-processing latitude
Lens24mm / 35mm primeZoom lens acceptable but tape the zoom ring locked
Aperturef/8–f/11Deep enough DoF (~0.8m at 1m subject distance), minimal distortion
ISO100–400Tripod allows locking ISO 100
Shutter≥1/125s (handheld) / 1/30s (tripod)Slower requires stabilizer
White balanceManual 5500K or custom gray card calibrationNever Auto
FocusMF + single-point lockKeeps focal plane consistent across all shots
Image stabilizationOFFIS causes ±0.5px unpredictable shifts between frames; SfM misinterprets these

Why turn off stabilization? In-body stabilization (IBIS) compensates shake by physically moving the sensor — meaning each frame's sensor position differs slightly. SfM assumes the sensor is fixed on the lens optical axis. IS breaks this assumption. On a tripod, it's even worse: the IS system "can't find shake" and creates micro-vibrations of its own.

360° Camera (Insta360 X5 / Theta Z1)

360° cameras are efficiency powerhouses for large spaces. In April 2026, Insta360 partnered with Splatica to offer automated 360° video → 3DGS pipelines: 5-minute capture → model ready in hours.

SettingValueNotes
ModePhoto priority, video secondarySame CMOS: video is 1/4 pixels of photo
FormatJPEG + DNG (if supported)DNG for post-processing headroom
HDRAlways on360° cameras have tight exposure latitude (~11 stops per eye)
White balanceManual 3200K (warm indoor) / 5500K (outdoor)Auto causes color shifts at stitch seams
ISO100–400Don't exceed 800; X5's 1/1.3" sensor shows visible noise above ISO 800
SpacingHorizontal 1.5–2m, vertical 3 layersWaist / chest / head height
Subject distance≥1.2m safe distanceToo close triggers fisheye distortion artifacts
Self-occlusionStitch seam aimed at your faceLets algorithm ignore you

Critical workflow: .insv → JPEG export

Never feed raw .insv files directly to training software. Export through Insta360 Studio first:

  1. Open Insta360 Studio → select .insv files

  2. Export settings: maximum resolution (X5 reaches 8192×4096)

  3. Disable FlowState stabilization (stabilization alters inter-frame geometry)

  4. Enable Direction Lock

  5. Enable PureShot / Clarity Enhancement

  6. Export as JPEG (or DNG for color grading)

Training failure rate with raw .insv exceeds 80%. After export, use a CubeMap split script to cut into 6 faces, then feed to COLMAP.

Action Camera (GoPro Hero 13 / Insta360 Ace Pro 2 / DJI Action 5)

Action cameras are the best value for outdoor work. The triple-mount rig — three cameras on one pole, voice-synced trigger — captures museums and galleries an order of magnitude faster.

SettingValueNotes
Mode4K 60fps SDRDon't shoot HDR video (HLG/Dolby Vision color pipeline is complex, training software incompatible)
Lens1× standard wide0.5× fisheye distortion too heavy for COLMAP correction
StabilizationOFFSame reason — stabilization makes inter-frame displacement uncontrollable
SpeedNormal speed recordingExtract every Nth frame with FFmpeg in post
White balanceManual lock (outdoor 5500K / indoor 4000K)Action camera Auto WB jumps severely
Low lightSwitch to photo modeVideo in low light drops shutter to 1/30s → severe motion blur

Triple-mount details: A 2m carbon fiber pole with GoPros at top, middle, and bottom. Use GoPro Labs firmware "voice trigger" for synchronized start. Three cameras recording simultaneously = continuous three-height coverage in a single pass.

Drone (DJI Mini 4 Pro / Mavic 3 Pro / Matrice 4E)

SettingValueNotes
ModeM RAW + JPGJPG feeds SfM directly; RAW for post backup
LensMain 1×Don't use telephoto (shallow DoF, narrow FoV, low coverage efficiency)
Flight modeSmart Orbit / Litchi waypoint planningAvoid hand-flying jitter; Litchi enables precise flight paths
SpeedSlow and constant <5m/sPrevent motion blur (5m/s + 1/500s shutter = 1cm displacement/frame)
AltitudeThree layers (look up 30° / level 0° / look down -45°)Single layer cannot reconstruct the top
OverlapForward 80%, lateral 70%Set in DJI Terra / Litchi flight planning
TimeOvercast / 2 hours after sunriseAvoid noon overhead light
Distance≥30m from subjectToo close = heavy distortion + safety risk

DJI Terra V5.2 workflow: Since July 2025, DJI Terra natively supports 3DGS output. Flight planning → autonomous flight → photo transfer → one-click .ply generation. For building exterior projects, this is currently the most hands-off end-to-end solution.

Small Spaces: The "Hard Nut" — No Room to Back Up

Small spaces (10–30㎡, like a single room) are 3DGS's hard nut. The reasons: you're only 1m from the wall, the frame is filled with featureless white wall; no continuous objects provide feature points; lens field of view isn't wide enough, each frame covers too little.

figure

The solution: three-layer shooting + loop path.

Three-Layer Shooting

Switch between three heights to ensure no vertical information gaps:

HeightValueLens direction
High1.9–2m (raised overhead)Level or slightly down 10°
MidChest height ~1.3mLevel
LowBent over ~0.7mDown 30°

Transitions between layers still need 5–6 gradual frames — never jump heights abruptly.

Loop Path Pattern

text
┌─────────────────────┐ │ ←──── Start ──────
│ │ ↑


Outer loop: walk along walls │ │
┌───────┐

│ │ │

←─→



Inner loop: orbit central object │ │




│ │ │
└───────┘


Key objects: extra 3 orbits │ ↓

│ │
───── End ─────→
│ └─────────────────────┘

Outer loop follows walls (facing opposite wall), inner loop orbits central object (facing inward), key objects get 3 extra orbits (high/mid/low).

Three Traps in Small Spaces

Trap 1: Macro trigger. Phone cameras switch to macro mode when <30cm from an object — only a small area is sharp, background completely blurred. Algorithm fails immediately. Maintain 30–50cm safe distance; step back and use 0.5× ultrawide rather than getting close.

Trap 2: Rotating in place. Standing against a wall and spinning 360° ≠ full coverage. The algorithm sees only "a panorama from one point" with zero depth. You must have a translational trajectory.

Trap 3: Featureless white walls. 3DGS's worst enemy. Three rescue techniques:

• Step back: Include ceiling, floor, and wall in one frame — give the algorithm 3 reference planes

• Shoot at an angle: Photograph walls obliquely so edge lines enter the frame

• Add texture: In your own space, hang a painting or poster as a feature anchor point

Large Spaces: The Opposite Problem — Insufficient Coverage

Large spaces (buildings, campuses, city blocks) have the opposite problem — you can never cover everything in one pass. 500 shots sounds like a lot, but spread across all facades of a three-story building, each angle gets only 2–3 shots — far from enough.

The solution: grid segmentation + multiple flights.

figure

Grid Segmentation Method

  1. Divide the large space into 30×30m grid cells

  2. Each cell gets a complete three-layer pass (ground 3 layers + drone 3 layers)

  3. Leave 30% overlap between cells for SfM stitching

  4. Each cell: estimated 80–150 shots (ground) + 50–100 shots (drone)

A 100×100m building complex requires approximately 9 cells × 150 = 1350 ground shots + 9 × 80 = 720 drone shots, totaling ~2000 images. Complete over 2–3 days (same time window each day to ensure lighting consistency).

Handling Moving Objects

3DGS assumes the scene is static. Anything that moves between two frames becomes a chaotic mess of floater artifacts.

ObjectHandling
Pedestrians, petsWait for them to leave. Or mask out in post
Wind-blown leaves, flagsChoose calm conditions, or actively avoid these angles
TV screensTurn off (screen content changes every frame)
Clocks (moving hands)Remove, or mask in post-training
Running water, fountainsKeep for aesthetic "silk" effect; turn off for precision
Construction vehicles, workersShoot during off-hours

figure

Next Steps

• Finished shooting, how to organize files → Enter 04-Dataset Organization & Archiving

• Shot video, how to extract frames → Enter 05-Video Frame Extraction & Quality Filtering

• Photo colors inconsistent → Enter 06-Color Grading & White Balance Unification

• ← Previous chapter: 02-Site Recon & Shot Planning