CV Chapter 3 Segmentation
Questions about the lecture 'Computer Vision' of the RWTH Aachen Chapter 3 Segmentation
Questions about the lecture 'Computer Vision' of the RWTH Aachen Chapter 3 Segmentation
Kartei Details
Karten | 53 |
---|---|
Sprache | English |
Kategorie | Informatik |
Stufe | Universität |
Erstellt / Aktualisiert | 04.02.2017 / 19.02.2017 |
Weblink |
https://card2brain.ch/box/20170204_cv_chapter_3_segmentation
|
Einbinden |
<iframe src="https://card2brain.ch/box/20170204_cv_chapter_3_segmentation/embed" width="780" height="150" scrolling="no" frameborder="0"></iframe>
|
What is the motivation of segmentation? [4]
1. Image regions
2. Video frames into shots
3. Figure-ground
4. Object-level
What are the basic questions of segmentation? [2]
1. What things should be grouped?
2. What cues indicate groups?
What are the characteristics of the Gestalt school? [3]
1. “Grouping key to visual perception”
2. Elements of collection can have properties resulting from relations
3. “Whole is greater than sum of its parts”
What are the Gestalt factors? [8]
1. Not grouped // . . . .
2. Proximity // .. .. .. ..
3. Similarity// ***..***
4. Common fate and region
5. Parallelism // << || >>
6. Symmetry // >< || <>
7. Continuity// Edges
8. Closure // o <>
What are difficulties regarding Gestalt factors? [2]
1. Difficult for algorithms
2. Occlusion covers above factors
What is the basic assumption for image segmentation?
Group of pixel that belong together
What are the characteristics of grouping by pixel intensity? [3]
1. Find representative centers and map pixels to nearest
2. Center minimizing SSD Sum|p-c|²
3. Center → Membership and Membership → Center
What is the definition of K-Means algorithm? [3]
1. Randomly initialize k cluster center
2. Determine points for centers
3. If center is false for points update and repeat
What are the issues of K-Means? [2]
1. Will find local but not always global minimum
2. Choose next proportional to |p-c|² till k centers with error O(log k)*optimal
What are the advantages of K-Means? [2]
1. Simple and fast
2. Converges to local minimum within cluster squared error
What are the disadvantages of K-Means? [4]
1. Which k? // NP-hard for even k=2
2. Sensitive to initial centers and outliers
3. Detects only convex clusters
4. Assuming means
What are possible feature spaces? [3]
1. 1D intensity similarity // Extended by position
2. 3D color similarity
3. 24D texture similarity // Possibilities from black to white
What are the basic questions of probabilistic clustering? [2]
1. What is the probability that point p belongs to cluster c?
2. What is the shape of the cluster c?
What is the basic assumption for generative model?
Assume that pixel are generated by continuous function
What are the characteristics of mixture of gaussians (MoG)? [4]
1. K Gaussian blobs with means mye_j cov matrices Sigma_j and dim D
2. p(x|theta_j) = 1/2pi^(D/2)*|Sigma_j|^(1/2) * exp(-1/2 *(x-mye_)^T *Sigma_j^-1 *(x-mye_j) )
3. Select blob j with probability pi_j
4. Likelihood of x is weighted mixture of Gaussians p(x|theta) = Sigma_j^K pi_j(x|theta_j)
What are the characteristics of expectation maximization (EM)? [3]
1. Maximize p(data|theta) = Prod_n^N p(x_n|theta)
2. Current blobs → Ownership // Expectation
3. Ownership probabilities → Update blobs // Maximization
What are the characteristics of expectation?
gamma_j(xn_) ← value_j / Sigma_k^K value_k with value_j =pi_j * N(x_n|mye_j,Sigma_j)
What are the characteristics of maximization? [4]
1. N_j ← Sum_n^N gamma_j(x_n)
2. pi_j^new ← N_j/N
3. mye_j^new ← 1/N_j * Sigma_n^N gamma_j(x_n)*x_n
4. Sigma_j^new ← 1/N_j * Sigma_n^N gamma_j(x_n)*(x_n-mye_j^new)^(T+1)
What is expectation maximization useful for? [3]
1. Any clustering problem // Any model estimation problem
2. Missing data problem and finding outliers
3. segmentation problems color motion fore/background
What are the advantages of expectation maximization? [3]
1. Soft assignments between data and cluster
2. Generative model can predict novel points
3. Compact storage
What are the disadvantages of expectation maximization? [3]
1. Initialization // With some k-means iterations
2. Unknown number of components // Solution with model selection (AIC, BIC), Dirichlet process mixture
3. Which generative model?
What is the basic idea of mean-shift segmentation? [2]
1. Modes in histogram are local maximum of the density
2. Cluster is set of points in attraction of one mode
What is the mean-shift segmentation algorithm? [3]
1. Initialize random seed and window W
2. Mean of W Sigma_xinW x*H(x)
3. Shift W to mean and redo 2. till convergence
What are the characteristics of real modality analysis? [3]
1. Choose multiple windows
2. Run mean-shift segmentations parallel
3. Label the traversed points
Which speedups for mean-shift segmentation? [2]
1. Label all points choose in radius r
2. Label all points in search path
What are the advantages of mean-shift segmentation? [4]
1. Model-free, no prior shape assumption
2. Only one parameter window size
3. Finds number of modes
4. Robust to outliers
What are the disadvantages of mean-shift segmentation? [2]
1. Output depends on not trivial window size
2. Expensive (2s/image)
What is the idea of markow random fields (MRF)? [3]
1. Rich probabilistic models for images
2. Lear local effects and get global effects out
3. Goal is to infer optimal labeling of MRF
What are the components of markov random fields (MRF)? [4]
1. Observed evidence on y-plane the image
2. Image scene compatibility function Phi(xi,yi)
3. Hidden true states on x-plane the scene
4. Neighborhood relations or scene scene compatibility function Psi(xi,xj)
What is the definition of the network joint probability p(x,y) of MRF?
Network joint probability p(x,y)=Prod_i Phi(xi,yi)* Prod_ij Psi(xi,xj)
What is the alternative of maximizing the network joint probability?
Minimizing -log p(x,y)=-Sum_i log(Phi(xi,yi)) - Sum_ij log(Psi(xi,xj))
What is the definition of the energy E(x,y) of MRF?
E(x,y) = Sum_i phi(xi,yi)) + Sum_ij psi(xi,xj)
What are the potentials of the energy of MRF? [2]
1. Unary or single-node phi encoding local info // membership of pixel to a class
2. Pairwise psi encoding neighborhood info
Which possibilities exist to set the potentials of the energy of MRF? [2]
1. Color modeled with MoG
2. Edge with contrast sensitive Potts model
Which inference algorithms for MRF exist? [6]
1. Gibbs sampling
2. Simulated annealing
3. Iterated conditional modes (ICM)
4. Variational methods
5. Belief propagation
6. Graph cuts
What are the characteristics of graph cuts for MRF? [4]
1. Popular fast tool for certain class of energy functions
2. Convert MRF into source-sink graph (SSG)
3. Minimum cost cut with max-flow/min-cut in polynomial time
4. Only minimizes submodular binary energies (Boros and Hummer 2002)
What is the definition of submodular of graph cut for MRF? [3]
1. E(s,s)+E(t,t)\leqE(s,t)+E(t,s)
2. Discrete equivalent to convexity
3. Every local is global minimum
What are the formulas of graph cut for MRF? [2]
1. wij = exp(- Diij/2*sigma²)
2. E(x,y) = Sum_i phi_i(xi) + Sum_ij wij * delta(xi!=xj)
What holds for the graph cut algorithm for MRF? [4]
1. Expect phi_i(s or t) ~ exp(-|Ii-(Is or It)|² /2*sigma²)
2. Is It are given expected intensities and can be re-estimated
3. Hard constrains are not required in general
4. Regional bias based on any intensity model with phi_i(Li)=-log p(Ii|Li) and Li s or t
What is the definition of the s-t-Mincut problem? [4]
1. Graph with vertices edges and costs
2. st-cut divides nodes between source and sink
3. Cost is sum of edges going from source to sink region
4. st-mincut has minimal cost